Advanced Warning! Voorrang! This post is highly non-linear and may cause headache!
Our group decided that we should participate in SPIE Photonics with some of our extra auxilary work. Hence, I ended up visiting the 5-day conference last week. The topics in this particular issue spanned from metamaterials and nanophotonics, via photonic crystals and devices, all the way to silicon photonics and integrated circuits, semiconductor lasers and quantum technologies. I was presenting a topic on noise in various CMOS TDI image sensor architectures, which ended-up in the photonic integrated circuits session. Weird, you might think, but the CMOS world and emerging silicon photonic technologies are very very close. What's currently missing is a good glue between both fields. Once we start jumping over this gap (maybe a timespan of 5-10 years???), we should hopefully see a boom in cheap, compact, optical spectrum analysis, microfluidic manipulation, gas sensing, molecule detection, chip-level molecule separation, multispectral imagers, photoacoustic sensing, terahertz imaging, quantum encryption and who knows what else applied to the real world. Hence, this visit was extremely useful to me. Useful, in a sense that I managed to get a live insight of the neighbouring to the electronic and imaging fields. I am offering you some of the impressions I managed to build and remember during these 5 busy days. [off-topic] Here is the place to mention that not all the presentations and presented work in this conference was brilliant. Unfortunately, I saw quite a number of striking talks which I would skip here. [/off-topic]
I kicked off the conference by listening a presentation about "Microstructured (micropyramid) IR hybrid detector design in GaAs", presented by a Slovenian group from Ljubljana. I wrote a few longer-ish comments about this talk, however I decided to comment them (though they are still kept in the source as comments). Nevertheless, here is the place to mention that I learned about black silicon - a semiconductor material formed by arrgessive reactive ion etching engraving tiny needles on the surface of silicon. This causes an increased absorption of light due to the increased Si surface. Because of the indirect bandgap nature of silicon we can achieve much higher IR sensitivity not only due to the increased area but also because of other second order effects such as reduced frensel reflection on the surface of the crystal. In the case of the presented work, frensel reflections were reduced thanks to the pyramidal shape of the detector. Hmmm, I am trying to make an analogy between acoustic anecoic chamber wall geometry and the presented by the Slovenian group photodiode design with respect to reflections...
Another intriguing work presented "An integrated SOI tunable polarization controller". The structure was formed by SiN waveguides, which fed light to three polarization rotators and three tuneable polarization phase shifters. By passing current through resistors placed only on the waveguides (forming the polarization phase shifters), they are able to precisely tune the wave phase. The light beam after each waveguide is fed to phase tuneable polarization rotators before being fed out. The group was using three polarization rotators and "heat-able" waveguides. As expected, the step response of such a (heat-controlled) system is very slow, however it is still applicable to imaging or coherent optical communication systems.
A group from IMEC presented "Density controlled nanophotonic waveguide gratings for efficient on-chip out-coupling in the near field". The origin of the problem they are trying to solve stems from holographic imaging of biological cells. Holograpgic imaging requires an accurate light point source, which lights the cells passing through a transparent microfluidic channel. At the back end of the channel a 2D camera (sensor) collects the scattered interference pattern caused by the bio-cell. Here is a crude sketch of their setup.
Apparently "standard" waveguides are not very easy to design for a wide range of wavelengths which can still act as accurate point sources. IMEC showed a methodology for creating a point source using a specifically optimized grid of blocking rods in a SiN medium to create pointsource which is accurate and works for a wide wavelenght bandwidth.
A japanese group presented "High-accuracy absolute distance measurement with a mode-resolved optical frequency comb", this talk somehow lost me completely, however, the whole point of their complex and expensive setup was to measure relatively long distances (~25-50m) with extremely high accuracy (um). Here is the place to mention that this work does not use the Time of Flight concept. Anyway, I learned some stuff about optical frequency combs and a methodology to generate a specific spectral fingerprint using light beam pulse trains in the time domain. Things in wave optics do really repeat very elegantly with RF electronics. The presenter mentioned the chain of terms - vernier light matching spectral counting, which immediately made me make the analogy with vernier based TDCs.
The next day started with a hot topics plenary session. John Dudley - the one coining the IYL initiative, gave an inspiring talk entitled "Lighting the Future of Photonics: the Legacy of the International Year of Light". He reviewed the accomplishments of the IYL and showed the truly remarkable number of educational events triggered by the outreach of IYL.
Cesar Misas from Univerisity of Jena presented "Current Challenges and Perspectives in High Power Fiber Lasers". He reviewed the challenges in acquiring high power from fiber lasers - the dominant of which seems to be laser mode instability? From the talk I managed to remember the following setup, which made their group capable of squeezing higher power out of the fiber medium.
By adding an acousto-optic beam steering AOB in front of the laser medium pump, controlled by an electronic feedback with the help of a photodiode, one is able to apply corrections to be pump power such that the output is stable and does not hover between two different discrete lasing modes.
Later, at the optical sensing session, Samuel Burri from EPFL presented a SPAD line scanner directly bonded into an FPGA "LinoSPAD: a time-resolved 256x1 CMOS SPAD line sensor system featuring 64 FPGA-based TDC channels running at up to 8.5 giga-events per second". A designed SPAD in 0.15um standard CMOS was formed into an array configuration with external passive quench resistors. The total 256 spad outputs were directly bonded out and fed into an FPGA. The latter contained 64 standard counter/phase difference TDCs and a sophisticated multiplexing and calibration network. Samuel presented nice histograms showing time of arrival, column mismatch and images in the form of 2D plots. My discussions with him hinted that the whole project was an entirely one man work which is quite remarkable.
A group from the University of Trento presented "Pixel-level continuous-time incremental sigma-delta A/D converter for THz sensors". This essentially presented a compact 1st order sigma-delta ADC. However, the application is a rather fancy one. Here is how I remember them showing their THz sense side:
Veronique Rochus from IMEC presented "Optical design of planar microlenses for enhanced pixel performance of CMOS imagers". Veronique started by introducing the main commercial microlense fabrication processing steps, identifying their pros and cons, which was right on the spot for me. Later, she presented a novel frensel microlense design for improved optical crosstalk using a three-layer fabrication steps. In addition, she presented an improved metamaterial-based frensel microlense designed by only two layers effectively utilizing the interference of light. The microlenses were tested on a standard CMOSIS CMV4000 sensor, it was also mentioned that further tests with more FSI/BSI sensors is on its way.
William Wardley from KCL displayed their work on "Large-area fabrication and characterization of UV regime metameterials manufactured using self-assembly techniques". Essentially the presentation was about a methodology of nanohole fabrication using chemical methods, avoiding expensive e-beam lithography or equivalent, while still maintaining a reasonable nanohole uniformity for the given application. Their nanohole array generation was based on anodization of aluminium and consecutive argon ion etching. They showed SEM images of arrays with separation from 60 to 200nm and a radius of about 10nm. One truly cheap method of sub-wavelength nanohole generation.
Brian Pogue from Thayer school of engineering presented "Cherenkov imaging for radiation dose and molecular sensing in vivo". A system developed for in vivo imaging and detection of cancer cells. The presented system employed a gated ICCD working in conjunction with standard CCD imagers, which are capturing alternating frames. The background lighting of the patients is also controlled by the gating controller such that it is turned on when the color CCD is taking a frame and turned off when the Intensified CCD is capturing the Cherenkov radiation. This allows for live viewing of the scene, as well as the simulataneous capture of the Cherenkov radiation. An interesting fact, the thresholds for Cherenkov radiation from water are \> 267 keV for electrons and \> 450MeV for protons.
Caeleste celebrated their 10th anniversary by organizing a workshop on the future of scientific and high-end image sensors. A few keynote speakers gave talks emphasizing on the past, present and future of these nieche imaging fields. The workshop was opened by (?) a guy from Aphesa which's name I cannot remember right now, the talk however, I can - CCDinosaurs vs CMOS image sensors. A nice review and food for thought of why the CCDinosaurs still roam the imaging planet and why the very same CCDinosaurs are big and stupid.
Karsten Sengebusch from Eureca gave a technical introduction to dithering in imaging with his presentation entitled "Prediction of the performance and image quality of CMOS image sensors". A large part of his talk looked like as if he was trying to convince the audience (or customers?) that they do not realy need a 12-bit ADC in their sensors, instead they could do with an 8, even 6-bits with added digital back-end dither. It was interesting for me to hear that Karsten is using dither with triangular distribution, which he claims provides best visual performance. Eureca are actively working on implementing real-time video dithering algorithms for their customers, which improve image quality after acquisition. A quick discussion with him about an early TDI dithering idea I had looked at, triggered a series of thoughts. Acquiring TDI lines with low quantization step and analog additive dither (or digital subtractive, or both) might not be a very bad idea after all, since the signal is tightly correlated for every line. It is just hidden within the quantized step, if we have a large set of stages, and if high speed imaging is needed (not low noise) dithering could possibly come handy in relaxing the ADC step requirements.
Benoit Dupont from Pyxalis gave an overview on the digital processors used for dual conversion gain combination in their HDR sensors, as well as timing generation. The title of the talk was "20-bit image sensors using dual processor architectures". He presented an architecture employing two RISC processors, one of them responsible for the DCG combine algorithms, tone mapping and lens shading correction, the other acted as a simple programmable sequencer for global chip-level timing generation. It was interesting to learn that Pyxalis can also sell the sequencers as an individual IP core to their customers.
Ajit Kumar from Caeleste presented a global shutter DCG pixel for HDR appications ("High Dynamic Range, shot noise limited Imagers with global shutter"). The talk was focused on a charge-domain GS pixel with a switchable FD capacitor for DCG operation. He presented some very early results and images of their pixel, which did not suffer from the typical ghosting effects as seen in rolling shutter HDR.
Jan Bosiers from Teledyne gave an overview of "Wafer-scale CMOS Imagers for medical X-ray imaging". He showed over 20+ different wafer-scale imagers designed by various groups around the world. An interesting fact - some of these large wafer-scale sensors employ additionl dosimetric pixels scattered around the photo array. These dosimetric pixels have a different structure and readout. The dosimetric information from the latter is used for calibration of the sensitivity and integration times of both the direct or indirect conversion imagers. According to Jan, there might be another 5/10+ years until organic photoconductive films reach the required resolution and performance of silicon detectors for (e.g.) dental applications.
Gert Finger from the IR instrumentation group at ESO presented "Large format and high speed sub-electron noise sensors for ground-based astronomy". A mind-blowind talk, showing a huge amount of imager projects used in the European Southern Observatory in Chile. Large ground-based telescopes require very sensitive large format photoarrays because of the focal planes of the telescopes and science instruments they use. The used materials in their project are by far not standard CMOS. Instead, the most promising used material I managed to remember was HgCdTe working at cryogenic temperatures. Their wavefront sensors also use avalanche multiplication detectors, by far not standard CMOS. He reported typical dark noise levels of 1 electron per 20 minutes - absolutely outstanding! It is worthless to mention that such noise levels are only achieved by operation in cryogenic temperatures. Gert mentioned that the newest sensors used in their Hawaii telescope arrays have (surprisingly) lower performance than the ones installed back in 2004.
Nick Nelms from the opto-electronics division at ESA gave an insight on ESA's space imaging projects with the talk entitled "High-performance image sensors in space, the shape of things to come". A highlight that cought my attention - ESA is actively trying to reduce the cost of missions (this includes the imaging field). However, on the other hand, ESA is also actively trying to use European vendors, such that the public money used in these missions are re-distributed back in Europe and not leaking out to overseas companies.
At the Photonic integrated circuits section, a group from the University of Gent presented "CMOS compatible SiN spectrometer for lab on a chip spectral sensing". The talk reviewed the a few key methodologies for spectral analysis on a chip. I managed to take a snapshot on the classification the authors made:
- arrayed waveguide grating using SiN photonic nanowires
- planar concave grating
-on-chip statuonary fourier transform spectrometry
- vibrational spectroscopy Interestingly, they missed (or I wasn't paying attention) the crude hyperspectral image sensor approach, which uses multiple color filters with different bandwidth.
I also noted a few presentations from the molecular sensors field. Anita Rogacs from HP labs in Palo Alto presented HP's efforts "Towards sensitive, low-cost, and field-deployable spectral analysis using SERS and flat optical components". As well as "Biosensors based on Si3N4 asymmetric Mach-Zender interferometers" from Tatevik Chalyan, Univ. Trento. Both projects aim the design of a field-deployable milk analyzer, used for early detection of toxins in milk caused by rotten corn/cow food. There is a large difference in both approaches and execution, but I guess that a few PhD students at an university can hardly compete with a dinosaur like HP Labs.
One of the last presentations in the biosensors session I listened to was that of Kristelle Robin from UCL on "Highly sensitive detection using miroring resonator and nanopores". It was clear that Kristelle presented a solely own project work which aimed at designing a sensor which can detect and hold a single molecule in place using an etched nanopore in a ring resonator. Complete measurements and characterization of the chip was presented showing promising detection results. Apparently currently this is sort-of possible with the help of nanopipettes - a truly scaled version of ordinary pipettes. It was mentioned that it was a pain to work with these, which is actually the driving force towards new designs.
To conclude, after a discussion and some argues with one of my colleagues that "the current scientific generation is a "consumer" one, and does not produce anything" . I would say that SPIE Photonics is actually a very good example, showing, that a lot of innovation is happening in the applied physics fields and this will continue to grow, no matter what.
Last, I am leaving my e-notes made during the conference, a crude equivalent to a post tag system which this site still lacks.
- microstructured (micropyramid) IR hybrid detector - responsivity 4 mA/W ??? - black silicon - micropyramid structured IR detector - SOI Integrated tuneable polatrization controller - thermally controller thermal inertial effects - low-loss CMOS copper plasmonic waveguides at the nanoscale - russian girl walking in the middle of talk, then asking stupid question at the end which the author explained in the beginng, ends by saying that the author was stupid...- light waveguides ~ 100nm order are hard to manufacture - plasmonics for optical interconnec - the promise of plasmonics (scientific american) from Dec 2014 - low-loss copper waveguides - plasmonic mode dispersion in nanoscale copper wavelength - density controller nanophotonic gratings for efficient on-chip out-coupling in the near field - on-chip microfluidic cell sorter - holographic imaging camera (scattered interference pattern) - high-accuracy absolute distance measurement with a mode-resolved optical frequency comb - optical frequency comb made with pulse train in the time domain - spectral interferometry - virtually imaged phase array - vernier light matching special counts - comb distance interferometry for our 50um - frequency comb-based depth imaging assisted by a low-coherent optical interferometer - OFC optical frequency comb - RF comb - Hadamard coding - Ghost imaging - Single pixel camera - Fiber lasers - high power, using feedback frequency lock loop - coherent addition of 4 lasers combined into one beam - Optical design of planar microlenses for enhanced pixel performance of CMOS imagers - metamaterial frensel lens - THz antenna mosfet - RING imaging Cherenkoc radiation (RICH radiator) - Transforming electromagentic reality to enhance Cherenkov radiation - Large area fabrication and characterization of UV regime metamaterials manufactured using self-alignment technologies [9883-19] - nanohole arrays coupling to grating - sub-wavelength nanoholes can act as metamaterial - separation 60nm to 200nm - radius 10nm to separation - nanohole array generation by anodisation and post-etching (argon ion etching) - Plasmonic nano-antennas - single-element plasmonic antenna, high-index dielectric nanoantenna - Optoacoustic imaging Alexander Graham Bell, Cambridge Vision Labs - Brain mapping project, BRAIN project - Seems like semiconductor + photonics is reserved for electrical eng conferences - Cherenkov imaging for radiation dose and molecular sensing in vivo - Threshold for Cherenkov imaging: - >267 keV electrons - >450 MeV protons - Emission properties of Cherenkov - avg peak intensities 100nW/cm2 to 1mW/cm2 - Usage of gated ICCD (intensified CCD), they do pulsed background light and cherenkov integration when background light is off - CCDinosaurs roam the imaging planet. CCDinosaurs are big and stupid - Dose sensing pixels within large are X-ray wafer scale sensors - Astronomical measurements - 1 photon capturing every 20 seconds - was at the LAST ATOM (LE DERNIER ATOME) - CMOS compatible SiN spectrometer for lab on a chip spectral sensing - Silicon Nitride SiN photonic integrated circuits - Silicon nitride opn chip spectrometer - arrayed waveguide grating SiN photonic wire - Planar concave grating - On-chip stationary Fourier Transform Spectrometer - Vibrational Spectroscopy - Silicon nitride photonics ICs are (sort-of) compatible with CMOS fabs - On-chip fluorescence excitation and collection by focusing grating couplers - Focusing gratinc couplers - 3D light field camera with microlense array for high resolution PIV
Both Cadence's Virtuoso and Mentor's Architect and Calibre family of tools require old linux infrastructure. The supported as of 2016 is usually RHEL 5 or 6. The problem is, I don't want to run RHEL at home, as it acts as a time-machine flying me 10 years into the past. It's a pain, although, a quite stable one.
Here's some quick insturctions on how you can prevent the following error messages and launch Virtuoso or Calibre on newer systems.
"Invalid Linux major version" WARNING: HOST gyrator DOES NOT APPEAR TO BE A CADENCE SUPPORTED LINUX CONFIGURATION.
The strategy for this fix is very simple, but yet I am including it here as I've been asked by my group for help about this way too often now. Both design environments, when launched, use bash scripts to identify the running environment and invoke their corresponding binary executables.
In the case with Mentor's Calibre tools, the preloaded skill in Virtuoso looks for and invokes: "./pkgs/calibre_base.aoi/bin/calibre_vco". The latter executes some checks on the OS before launching Calibre's aoi executable. We look for the following if statement and hence, modify the case conditions.
... if test -r /etc/redhat-release then major_rev=`grep release /etc/redhat-release 2>/dev/null \ | sed -e 's/.*release *//' \ | sed -e 's/\..*//' \ | sed -e 's/ .*//'` case "$major_rev" in 5) echo ixl ;; [6-9]) echo aoi ;; *) echo "ERROR: Invalid Linux major version '$major_rev'" >&2 echo exit 1 ;; esac . . .... and so on
You can also remove the exit flags, or even forge the whole if branch (althoug I would not recommend it).
The same procedure applies to Virtuoso. You start by locating the staring korn shell with e.g. "which virtuoso", then you will find out that the script invokes something like "{ROOT_DIR}/share/bin/.cdnWrapper_core". Then by looking at the wrapper, you will find out that it looks for the "${oabindir}/sysname" script. By opening the last you find out that it performs some checks using stadard shell commands (uname -a).
Then you look for the following if branches:
... case $machine in ia64 ) sysname="linux_rhas21_ia64$compiler"; sysnames="$sysname $sysnames";; *86 | *86_64 ) case $version in 2.4.* ) compiler="_gcc411" sysname="linux_rhel30$compiler"; sysnames="$sysname $sysnames";; 2.6.[0-9]-* ) compiler="_gcc44x" sysname="linux_rhel40$compiler"; sysnames="$sysname $sysnames";; 2.6.* | 4.1.*) compiler="_gcc44x" sysname="linux_rhel50$compiler"; sysnames="$sysname linux_rhel40$compiler $sysnames";; * ) check_global;; esac;; ... and so on ...
And here you can modify the conditions to the wanted kernel. In my case, just adding an additional OR | operator with my OS's kernel does the job.
The same applies for PVS, Spectre and the whole family of tools. I should warn however that using such CAD tools with an unsupported platform may lead to unexpexted results. In my case, I have noticed that such a new kernel, negatively affects the stability when performing SPICE simulations with Spectre/Ultrasim. Getting fake SPICE results without any error messages is the least that you want from a circuit simulator.
And so, finally, things are wrapping-up and my 1-year chip-design "military service" is coming to an end. But hey, where am I going without this little fella :)
Check out some more here.
Have you ever had this irritating issue with multiple unsaved schematic diagrams lower in the hierarchy preventing you from netlisting? Well, now we have a solution. Just bind that function to a handy key and blow the whistle when in trouble :)
; Checks and saves all cells in a given library name, ; Initial version A - Deyan Levski, 19.08.2015 ; ; procedure( checkAndSave(libName) notSave = '() lib = ddGetObj( libName ) println(libName) println(lib~>name) cellList = lib~>cells~>name foreach( cell cellList cellview = dbOpenCellViewByType( libName cell "schematic" "" "a" ) if(cellview != nil then schCheck(cellview) dbSave(cellview) dbClose(cellview) println(cell) println("saved") else notSave = append1(notSave cell) ) ) println("Cells not checked:") ) ; procedure
A hint: you can use " libName = geGetEditCellView()~>libName " to automatically fetch the library name you are currently in.
I thought that it might be a good idea to share the visuals from ISSCC with you and give short comments on some of the works I found intriguing from the CIS section. Please disregard my punctuation and language, this initially used to be a brief e-mail to my coworkers, however, I thought, "Why not get a free post out of it?" :) Unfortunately I am not allowed to upload any of the materials I sent them, thus you will have to refer to the original papers at IEEE XPlore.
Some new imaging materials made it to the CIS section this year, in particular there were two intriguing papers from Panasonic presenting an organic photoconductive global shutter imager with 120dB DR and speed of 60fps at almost 1MP with quite low read noise too!!! Also, note - they report a 600ke- FW capacity for a 6x6um pixel - impressive! There was a live demonstration of the imager, which looked rather sneaky as they used a very bright light (car headlamp) for illumination, which suggests low quantum efficiency of the sensor. Unfortunately, Panasonic did not provide any further comments on their work, apart from what is there in the slides. Some people asked direct questions on QE and process - answers were simply "I don't know".
The original paper is entitled as: An Over 120dB Simultaneous-Capture Wide-Dynamic-Range 1.6e- Ultra-Low-Reset-Noise Organic-Photoconductive-Film CMOS Image Sensor
Jan Bogaerts from CMOSIS presented an almost 400MP imager for airborne mapping, working at 1fps in 14-bit mode. Really impressive work from an engineering point of view. I managed to get hold of the sensor which has the size of a large smartphone, and one can clearly see the 6x3 pixel die blocks stitched together. The imager has no color filters, however the whole system uses 6 additional CCD sensors for color information extraction and a quite sophisticated optical stabilization system.
For more details refer to: 105x65mm2 391Mpixel CMOS Image Sensor with >78dB Dynamic Range for Airborne Mapping Applications
David Stoppa presented two works on SPAD imagers, one of which you can find in the visuals. The other presentation was given at a forum and have no electronic slides to show. The impressive point within these SPAD projects is that they report a read noise as low as 0.22 e-. This in terms allows them to measure photon shot noise and observe its poisson distribution! Theoretical physics works!
Paper title: A 64x64-Pixel Digital Silicon Photomultiplier Direct ToF Sensor with 100MPhotons/s/pixel Background Rejection and Imaging/Altimeter Mode with 0.14% Precision up to 6km for Spacecraft Navigation and Landing
TSMC, Toshiba and NHK presented works focusing on the ADC and analog signal processing in 3D stacked chips. Although, NHK and Toshiba present a refined version of their old cyclic ADC architecture, it is still worth noting that they report pure conversion time speeds of less than 0.9us, with relatively low energy per conversion. TSMC show detailed schematics of their Gray-coded ramp ADC, worth having a deeper investigation.
Paper titles accordingly: A 1.2e- Temporal Noise 3D-Stacked CMOS Image Sensor with Comparator-Based Multiple-Sampling PGA; A 1.1um 33Mpixel 240fps 3D-Stacked CMOS Image Sensor with 3-Stage Cyclic-Based Analog-to-Digital Converters
In addition TSMC's work features an interesting readout technique named "Negative Substrate Bias Readout" - In simple words: by biasing the substrate negatively they can use 1.5V supply for the readout and still get 3.3 volts for the pixel diode - thus saving power and not compromising SNR. This technique is enabled by the 3D stacking as the pixel array is a separate die, which can have a separate substrate bias.
Ref: A 1.5V 33Mpixel 3D-Stacked CMOS Image Sensor with Negative Substrate Bias
For more information and details around the presented works, I suggest having a further look at the digest materials here.