Secondary Electron Endpoint - FIB SEM User Group

68 downloads 118 Views 2MB Size Report
Mar 2, 2017 - •Design of tests targeting intended applications. •Practical ... Quantifiable testing of critical performance parameters ... Basic test of tool stability.
QUANTIFIABLE COMPARATIVE EVALUATION OF FIB/SEM INSTRUMENTS Valery Ray(1**), Joshua Taillon(2*), Lourdes Salamanca-Riba(2***) (1)

PBS&T, Methuen, MA; (2) University of Maryland, College Park, MD (*) Present address: Material Measurement Laboratory, NIST, Gaithersburg, MD (**) [email protected], (***) [email protected]

March 2, 2017

10th FIB SEM User Group Meeting, NIST Gaithersburg MD

Outline • Challenges of FIB/SEM equipment evaluation • Quantifiable comparative testing approach • Design of tests targeting intended applications • Practical Examples • Summary

March 2, 2017

10th FIB SEM User Group Meeting, NIST Gaithersburg MD

2

Challenges of Evaluating FIB/SEM Instruments • Complexity of translating application needs into

instrumentation requirements and evaluation criteria • There are no “bad” instruments out there • OEM engineers are highly skilled with demonstrations • Outcome of same operation for average user could be very different

• “Canned Demo” approach by OEMs • Designed to demonstrate strong sides

• Art of crafting specifications – “specsmanship” • Critical (for your application) performance parameters

could be “confidential” • Sometimes for a reason of not being known, defined, or ever tested March 2, 2017

10th FIB SEM User Group Meeting, NIST Gaithersburg MD

3

Quantifiable Comparative Testing Approach • Identify range of applications for intended usage • Translate application goals into instrumentation requirements

• Design comparative tests, define evaluation criteria • Test descriptions and samples to all vendors as early as possible

• Comprehensive evaluation based for intended use: • Quantifiable testing of critical performance parameters • Based on pre-defined evaluation criteria

• Applications demo • Overall performance in 3D applications, TEM lamella prep, etc…

• Two-day evaluation is reasonable to get all the data March 2, 2017

10th FIB SEM User Group Meeting, NIST Gaithersburg MD

4

Tests targeting intended applications • General Performance • Beam quality; System stability; Aperture repeatability

• Patterning • Beam placement; Etching fidelity; Beam drifts and shifts

• TEM lamella preparation • Throughput; Thickness uniformity; Ease of use; Automation; Endpoint

• FIB Tomography 3D slice-n-view • Unattended runtime; Image quality; Throughput; Ease of use; Drift

Correction; Focus Tracking; Slice thickness uniformity; EDS integration

• Imaging • SEM SE, SEM BSE, STEM BF, STEM DF, FIB SE, FIB SI….. March 2, 2017

10th FIB SEM User Group Meeting, NIST Gaithersburg MD

5

Samples for comparative evaluation • Performance Testing SiO2 optical flat, ~24nm evaporated Al coating, silver paint around perimeter Aperture Repeatability

• Application Testing Epoxy-impregnated Solid Electrolyte Fuel Cell (SOFC), 2-phase ceramic

Deposition Profiles

SiO2 Etching Profiles

TEM lift-out

Same sample(s) to all vendors, require return of test sample(s) for independent analysis March 2, 2017

10th FIB SEM User Group Meeting, NIST Gaithersburg MD

6

General Performance – Beam Quality • Basic test of ion beam quality: shape and homogeneity Aperture size increase

Aperture size increase

Inhomogeneous beam Aperture size increase

Single aperture problem

March 2, 2017

Aperture size increase

Dose increase

Systematic high-current problem

10th FIB SEM User Group Meeting, NIST Gaithersburg MD

7

General Performance – System Stability Typical system drift

• Basic test of tool stability HVAC failure

20 µm 1st pair

“Quite not bad”

1 µm

L-shaped single-line pattern and 21 beam burns on each arm with 1um offset along the line and 5 minutes delay between burns

20 µm March 2, 2017

10th FIB SEM User Group Meeting, NIST Gaithersburg MD

8

Critical Performance – Etch Placement & Fidelity 1pA

10pA

3pA

1.8 nC/µm2

0.6 nC/µm2

Gas

0.2 nC/µm2

1.8 nC/µm2

0.6 nC/µm2

Gas

0.2 nC/µm2

1.8 nC/µm2

0.2 nC/µm2

0.6 nC/µm2

1um

Gas

Shortest dwell time, -20% pixel overlap, x2 pattern repeats: (a) sputtering/GAE (XeF2) and (b) sputtering/depo (C, Pt, W) March 2, 2017

10th FIB SEM User Group Meeting, NIST Gaithersburg MD

9

Critical Performance – Etching Placement • Patterning where intended with/without gas injection Drift after aperture change Problem for automatic patterning

Shift due to gas problem for site-specific deposition and GAE

Aperture change shift problem for multiplecurrent patterning

Visible artifacts

Drift during line exposure

Expected performance

• C or Pt stripe e-beam deposited across lines, TEM lamella

prepared and STEM-imaged as part of application testing March 2, 2017

10th FIB SEM User Group Meeting, NIST Gaithersburg MD

10

Critical Performance – Etching Fidelity Al intact

No-Gas Narrowest cut defined by width of a tip of No-Gas etching profile No-Gas to GAE profile area ratio defines GAE enhancement March 2, 2017

Sidewall slope and No-Gas profile aspect ratio define polishing efficiency

Al intact

Al layer removal indicates beam tails damage to surface Al removed beam tails damage

GAE Al removed beam tails damage

No-Gas 10th FIB SEM User Group Meeting, NIST Gaithersburg MD

GAE 11

Application Testing – SOFC imaging • Side-by-side comparison of same sample imaging

SE

BSE

March 2, 2017

10th FIB SEM User Group Meeting, NIST Gaithersburg MD

12

Application Testing – 3D Reconstruction • Fix experimental

parameters between vendors: • Run overnight, if possible

• Results to evaluate: • Total running time (limited by stability) • Usable acquisition volume/hour • Acquired image quality • Output/ease of use of 3D visualization software March 2, 2017

Slice thickness

Detector settings

Image resolution

Dwell time

Example of vendor visualization output

10th FIB SEM User Group Meeting, NIST Gaithersburg MD

13

Summary • Quantifiable testing approach enables comparative

evaluation of FIB/SEM instruments by collecting performance data under controlled conditions • Careful sample preparation, thorough test design, and demo planning

• Seamless integration of performance tests with

applications demo facilitates comprehensive evaluation • providing OEMs opportunity to showcase strong features of the equipment • while allowing side-by-side comparison of critical performance parameters

• There are no “bad” tools, but nobody is perfect either • Interpret test results in context of realistic application requirements

March 2, 2017

10th FIB SEM User Group Meeting, NIST Gaithersburg MD

14

Suggest Documents