Kamis, 03 Oktober 2019

MARIA PREFER in electronic components traveling a few light years will use photons in future electronic equipment to replace the speed of the switching process and signal conditioner in the form of old electronic components such as silicon and germanium microcontrolers and nanocontrolers AMNIMARJESLOW GOVERMENT SUIT 91220017 Sū-kōnen idō suru denshi buhin wa, shōrai no denshi kiki de kōshi o shiyō shimasu 0209601001 LJBUS SPACE 21 forward 22 -24 ___ Thanks to Lord Jesus ... God who is the father of heaven who is in control with his light he guides our steps to calm water ___ Gen.Mac Tech Zone The Future is light on electronics component for ligh velocity at space ship








                                            Robots with human-like flesh, robotic vision and conversation skills via artificial intelligence will greet guests in the reception area of the Henn-na Hotel, planned to open this year in Nagasaki, Japan.

              

                            Electronics

Electronics comprises the physics, engineering, technology and applications that deal with the emission, flow and control of electrons in vacuum and matter.
Electronics is widely used in information processing, telecommunication, and signal processing. The ability of electronic devices to act as switches makes digital information-processing possible. Interconnection technologies such as circuit boards, electronics packaging technology, and other varied forms of communication infrastructure complete circuit functionality and transform the mixed electronic components into a regular working system, called an electronic system; examples are computers or control systems. An electronic system may be a component of another engineered system or a standalone device. As of 2018 most electronic devices use semiconductor components to perform electron control.
The identification of the electron in 1897, along with the invention of the vacuum tube, which could amplify and rectify small electrical signals, inaugurated the field of electronics and the electron age.
Commonly, electronic devices contain circuitry consisting primarily or exclusively of active semiconductors supplemented with passive elements; such a circuit is described as an electronic circuit. Electronics deals with electrical circuits that involve active electrical components such as vacuum tubes, transistors, diodes, integrated circuits, optoelectronics, and sensors, associated passive electrical components, and interconnection technologies.The nonlinear behaviour of active components and their ability to control electron flows makes amplification of weak signals possible.
Electrical and electromechanical science and technology deals with the generation, distribution, switching, storage, and conversion of electrical energy to and from other energy forms (using wires, motors, generators, batteries, switches, relays, transformers, resistors, and other passive components). This distinction started around 1906 with the invention by Lee De Forest of the triode, which made electrical amplification of weak radio signals and audio signals possible with a non-mechanical device. Until 1950, this field was called "radio technology" because its principal application was the design and theory of radio transmitters, receivers, and vacuum tubes.
The term "solid-state electronics" emerged after the first working transistor was invented by William Shockley, Walter Houser Brattain and John Bardeen at Bell Labs in 1947. The MOSFET (MOS transistor) was later invented by Mohamed Atalla and Dawon Kahng at Bell Labs in 1959. The MOSFET was the first truly compact transistor that could be miniaturised and mass-produced for a wide range of uses, revolutionizing the electronics industry, and playing a central role in the microelectronics revolution and Digital Revolution. The MOSFET has since become the basic element in most modern electronic equipment, and is the most widely used electronic device in the world.
The study of semiconductor devices and related technology is considered a branch of solid-state physics, whereas the design and construction of electronic circuits to solve practical problems come under electronics engineering. This article focuses on engineering aspects of electronics.

                                       
                                         Surface-mount electronic components

Branches of electronics

Electronics has branches as follows:
  1. Digital electronics
  2. Analogue electronics
  3. Microelectronics
  4. Circuit design
  5. Integrated circuits
  6. Power electronics
  7. Optoelectronics
  8. Semiconductor devices
  9. Embedded systems
  10. Audio electronics
  11. Telecommunications 
  12. control electronic on e- Money 
An electronic component is any physical entity in an electronic system used to affect the electrons or their associated fields in a manner consistent with the intended function of the electronic system. Components are generally intended to be connected together, usually by being soldered to a printed circuit board (PCB), to create an electronic circuit with a particular function (for example an amplifier, radio receiver, or oscillator). Components may be packaged singly, or in more complex groups as integrated circuits. Some common electronic components are capacitors, inductors, resistors, diodes, transistors, etc. Components are often categorized as active (e.g. transistors and thyristors) or passive (e.g. resistors, diodes, inductors and capacitors ) .

Types of circuits

Circuits and components can be divided into two groups: analog and digital. A particular device may consist of circuitry that has one or the other or a mix of the two types.

Analog circuits

Hitachi J100 adjustable frequency drive chassis
Most analog electronic appliances, such as radio receivers, are constructed from combinations of a few types of basic circuits. Analog circuits use a continuous range of voltage or current as opposed to discrete levels as in digital circuits.
The number of different analog circuits so far devised is huge, especially because a 'circuit' can be defined as anything from a single component, to systems containing thousands of components.
Analog circuits are sometimes called linear circuits although many non-linear effects are used in analog circuits such as mixers, modulators, etc. Good examples of analog circuits include vacuum tube and transistor amplifiers, operational amplifiers and oscillators.
One rarely finds modern circuits that are entirely analog. These days analog circuitry may use digital or even microprocessor techniques to improve performance. This type of circuit is usually called "mixed signal" rather than analog or digital.
Sometimes it may be difficult to differentiate between analog and digital circuits as they have elements of both linear and non-linear operation. An example is the comparator which takes in a continuous range of voltage but only outputs one of two levels as in a digital circuit. Similarly, an overdriven transistor amplifier can take on the characteristics of a controlled switch having essentially two levels of output. In fact, many digital circuits are actually implemented as variations of analog circuits similar to this example – after all, all aspects of the real physical world are essentially analog, so digital effects are only realized by constraining analog behavior.

Digital circuits

Digital circuits are electric circuits based on a number of discrete voltage levels. Digital circuits are the most common physical representation of Boolean algebra, and are the basis of all digital computers. To most engineers, the terms "digital circuit", "digital system" and "logic" are interchangeable in the context of digital circuits. Most digital circuits use a binary system with two voltage levels labeled "0" and "1". Often logic "0" will be a lower voltage and referred to as "Low" while logic "1" is referred to as "High". However, some systems use the reverse definition ("0" is "High") or are current based. Quite often the logic designer may reverse these definitions from one circuit to the next as he sees fit to facilitate his design. The definition of the levels as "0" or "1" is arbitrary.
Ternary (with three states) logic has been studied, and some prototype computers made.
Computers, electronic clocks, and programmable logic controllers (used to control industrial processes) are constructed of digital circuits. Digital signal processors are another example.
Building blocks:
Highly integrated devices:
Electronic noise is defined as unwanted disturbances superposed on a useful signal that tend to obscure its information content. Noise is not the same as signal distortion caused by a circuit. Noise is associated with all electronic circuits. Noise may be electromagnetically or thermally generated, which can be decreased by lowering the operating temperature of the circuit. Other types of noise, such as shot noise cannot be removed as they are due to limitations in physical properties.

Electronics theory

Mathematical methods are integral to the study of electronics. To become proficient in electronics it is also necessary to become proficient in the mathematics of circuit analysis.
Circuit analysis is the study of methods of solving generally linear systems for unknown variables such as the voltage at a certain node or the current through a certain branch of a network. A common analytical tool for this is the SPICE circuit simulator.
Also important to electronics is the study and understanding of electromagnetic field theory.

Electronics lab

Due to the complex nature of electronics theory, laboratory experimentation is an important part of the development of electronic devices. These experiments are used to test or verify the engineer's design and detect errors. Historically, electronics labs have consisted of electronics devices and equipment located in a physical space, although in more recent years the trend has been towards electronics lab simulation software, such as CircuitLogix, Multisim, and PSpice.

Computer aided design (CAD)

Today's electronics engineers have the ability to design circuits using premanufactured building blocks such as power supplies, semiconductors (i.e. semiconductor devices, such as transistors), and integrated circuits. Electronic design automation software programs include schematic capture programs and printed circuit board design programs. Popular names in the EDA software world are NI Multisim, Cadence (ORCAD), EAGLE PCB and Schematic, Mentor (PADS PCB and LOGIC Schematic), Altium (Protel), LabCentre Electronics (Proteus), gEDA, KiCad and many others.

Packaging methods

Many different methods of connecting components have been used over the years. For instance, early electronics often used point to point wiring with components attached to wooden breadboards to construct circuits. Cordwood construction and wire wrap were other methods used. Most modern day electronics now use printed circuit boards made of materials such as FR4, or the cheaper (and less hard-wearing) Synthetic Resin Bonded Paper (SRBP, also known as Paxoline/Paxolin (trade marks) and FR2) – characterised by its brown colour. Health and environmental concerns associated with electronics assembly have gained increased attention in recent years, especially for products destined to the European Union, with its Restriction of Hazardous Substances Directive (RoHS) and Waste Electrical and Electronic Equipment Directive (WEEE), which went into force in July 2006.

Electronic systems design

Electronic systems design deals with the multi-disciplinary design issues of complex electronic devices and systems, such as mobile phones and computers. The subject covers a broad spectrum, from the design and development of an electronic system (new product development) to assuring its proper function, service life and disposal. Electronic systems design is therefore the process of defining and developing complex electronic devices to satisfy specified requirements of the user.


Mounting Options

Electrical components are generally mounted in the following ways:
  • Through-hole
  • Surface Mount
  • Chassis Mount
  • LGA/BGA/PGA Socket
   ______________________________________________________________________________

            The future of electronics is light 

For the past four decades, the electronics industry has been driven by what is called “Moore’s Law,” which is not a law but more an axiom or observation. Effectively, it suggests that the electronic devices double in speed and capability about every two years. And indeed, every year tech companies come up with new, faster, smarter and better gadgets.
Specifically, Moore’s Law, as articulated by Intel cofounder Gordon Moore, is that “The number of transistors incorporated in a chip will approximately double every 24 months.” Transistors, tiny electrical switches, are the fundamental unit that drives all the electronic gadgets we can think of. As they get smaller, they also get faster and consume less electricity to operate.
In the technology world, one of the biggest questions of the 21st century is: How small can we make transistors? If there is a limit to how tiny they can get, we might reach a point at which we can no longer continue to make smaller, more powerful, more efficient devices. It’s an industry with more than US$200 billion in annual revenue in the U.S. alone. Might it stop growing?

Getting close to the limit

At the present, companies like Intel are mass-producing transistors 14 nanometers across – just 14 times wider than DNA molecules. They’re made of silicon, the second-most abundant material on our planet. Silicon’s atomic size is about 0.2 nanometers.
Today’s transistors are about 70 silicon atoms wide, so the possibility of making them even smaller is itself shrinking. We’re getting very close to the limit of how small we can make a transistor.
At present, transistors use electrical signals – electrons moving from one place to another – to communicate. But if we could use light, made up of photons, instead of electricity, we could make transistors even faster. My work, on finding ways to integrate light-based processing with existing chips, is part of that nascent effort.

Putting light inside a chip

A transistor has three parts; think of them as parts of a digital camera. First, information comes into the lens, analogous to a transistor’s source. Then it travels through a channel from the image sensor to the wires inside the camera. And lastly, the information is stored on the camera’s memory card, which is called a transistor’s “drain” – where the information ultimately ends up.


Light waves can have different frequencies. maxhurtz
Right now, all of that happens by moving electrons around. To substitute light as the medium, we actually need to move photons instead. Subatomic particles like electrons and photons travel in a wave motion, vibrating up and down even as they move in one direction. The length of each wave depends on what it’s traveling through.
In silicon, the most efficient wavelength for photons is 1.3 micrometers. This is very small – a human hair is around 100 micrometers across. But electrons in silicon are even smaller – with wavelengths 50 to 1,000 times shorter than photons.
This means the equipment to handle photons needs to be bigger than the electron-handling devices we have today. So it might seem like it would force us to build larger transistors, rather than smaller ones.
However, for two reasons, we could keep chips the same size and deliver more processing power, shrink chips while providing the same power, or, potentially both. First, a photonic chip needs only a few light sources, generating photons that can then be directed around the chip with very small lenses and mirrors.
And second, light is much faster than electrons. On average photons can travel about 20 times faster than electrons in a chip. That means computers that are 20 times faster, a speed increase that would take about 15 years to achieve with current technology.
Scientists have demonstrated progress toward photonic chips in recent years. A key challenge is making sure the new light-based chips can work with all the existing electronic chips. If we’re able to figure out how to do it – or even to use light-based transistors to enhance electronic ones – we could see significant performance improvement.

When can I get a light-based laptop or smartphone?

We still have some way to go before the first consumer device reaches the market, and progress takes time. The first transistor was made in the year 1907 using vacuum tubes, which were typically between one and six inches tall (on average 100 mm). By 1947, the current type of transistor – the one that’s now just 14 nanometers across – was invented and it was 40 micrometers long (about 3,000 times longer than the current one). And in 1971 the first commercial microprocessor (the powerhouse of any electronic gadget) was 1,000 times bigger than today’s when it was released.
The vast research efforts and the consequential evolution seen in the electronics industry are only starting in the photonic industry. As a result, current electronics can perform tasks that are far more complex than the best current photonic devices. But as research proceeds, light’s capability will catch up to, and ultimately surpass, electronics’ speeds. However long it takes to get there, the future of photonics is bright.



________________________________________________________________________________

     

                                Photon

The photon is a type of elementary particle. It is the quantum of the electromagnetic field including electromagnetic radiation such as light and radio waves, and the force carrier for the electromagnetic force (even when static via virtual particles). The invariant mass of the photon is zero; it always moves at the speed of light in a vacuum.
Like all elementary particles, photons are currently best explained by quantum mechanics and exhibit wave–particle duality, exhibiting properties of both waves and particles. For example, a single photon may be refracted by a lens and exhibit wave interference with itself, and it can behave as a particle with definite and finite measurable position or momentum, though not both at the same time as per Heisenberg's uncertainty principle. The photon's wave and quantum qualities are two observable aspects of a single phenomenon—they cannot be described by any mechanical model; a representation of this dual property of light that assumes certain points on the wavefront to be the seat of the energy is not possible. The quanta in a light wave are not spatially localized.
The modern concept of the photon was developed gradually by Albert Einstein in the early 20th century to explain experimental observations that did not fit the classical wave model of light. The benefit of the photon model is that it accounts for the frequency dependence of light's energy, and explains the ability of matter and electromagnetic radiation to be in thermal equilibrium. The photon model accounts for anomalous observations, including the properties of black-body radiation, that others (notably Max Planck) had tried to explain using semiclassical models. In that model, light is described by Maxwell's equations, but material objects emit and absorb light in quantized amounts (i.e., they change energy only by certain particular discrete amounts). Although these semiclassical models contributed to the development of quantum mechanics, many further experiments beginning with the phenomenon of Compton scattering of single photons by electrons, validated Einstein's hypothesis that light itself is quantized.[5] In December 1926, American physical chemist Gilbert N. Lewis coined the widely adopted name "photon" for these particles in a letter to Nature] After Arthur H. Compton won the Nobel Prize in 1927 for his scattering studies, most scientists accepted that light quanta have an independent existence, and the term "photon" was accepted.
In the Standard Model of particle physics, photons and other elementary particles are described as a necessary consequence of physical laws having a certain symmetry at every point in spacetime. The intrinsic properties of particles, such as charge, mass, and spin, are determined by this gauge symmetry. The photon concept has led to momentous advances in experimental and theoretical physics, including lasers, Bose–Einstein condensation, quantum field theory, and the probabilistic interpretation of quantum mechanics. It has been applied to photochemistry, high-resolution microscopy, and measurements of molecular distances. Recently, photons have been studied as elements of quantum computers, and for applications in optical imaging and optical communication such as quantum cryptography.

Einstein's light quantum

Unlike Planck, Einstein entertained the possibility that there might be actual physical quanta of light—what we now call photons. He noticed that a light quantum with energy proportional to its frequency would explain a number of troubling puzzles and paradoxes, including an unpublished law by Stokes, the ultraviolet catastrophe, and the photoelectric effect. Stokes's law said simply that the frequency of fluorescent light cannot be greater than the frequency of the light (usually ultraviolet) inducing it. Einstein eliminated the ultraviolet catastrophe by imagining a gas of photons behaving like a gas of electrons that he had previously considered. He was advised by a colleague to be careful how he wrote up this paper, in order to not challenge Planck, a powerful figure in physics, too directly, and indeed the warning was justified, as Planck never forgave him for writing it.  

               
             

                             Optics

                     title page of book

Optics is the branch of physics that studies the behaviour and properties of light, including its interactions with matter and the construction of instruments that use or detect it. Optics usually describes the behaviour of visible, ultraviolet, and infrared light. Because light is an electromagnetic wave, other forms of electromagnetic radiation such as X-rays, microwaves, and radio waves exhibit similar properties.
Most optical phenomena can be accounted for using the classical electromagnetic description of light. Complete electromagnetic descriptions of light are, however, often difficult to apply in practice. Practical optics is usually done using simplified models. The most common of these, geometric optics, treats light as a collection of rays that travel in straight lines and bend when they pass through or reflect from surfaces. Physical optics is a more comprehensive model of light, which includes wave effects such as diffraction and interference that cannot be accounted for in geometric optics. Historically, the ray-based model of light was developed first, followed by the wave model of light. Progress in electromagnetic theory in the 19th century led to the discovery that light waves were in fact electromagnetic radiation.
Some phenomena depend on the fact that light has both wave-like and particle-like properties. Explanation of these effects requires quantum mechanics. When considering light's particle-like properties, the light is modelled as a collection of particles called "photons". Quantum optics deals with the application of quantum mechanics to optical systems.
Optical science is relevant to and studied in many related disciplines including astronomy, various engineering fields, photography, and medicine (particularly ophthalmology and optometry). Practical applications of optics are found in a variety of technologies and everyday objects, including mirrors, lenses, telescopes, microscopes, lasers, and fibre optics.

                                            
                                     Optics includes study of dispersion of light.

Dispersion and scattering

Conceptual animation of light dispersion through a prism. High frequency (blue) light is deflected the most, and low frequency (red) the least.
Refractive processes take place in the physical optics limit, where the wavelength of light is similar to other distances, as a kind of scattering. The simplest type of scattering is Thomson scattering which occurs when electromagnetic waves are deflected by single particles. In the limit of Thomson scattering, in which the wavelike nature of light is evident, light is dispersed independent of the frequency, in contrast to Compton scattering which is frequency-dependent and strictly a quantum mechanical process, involving the nature of light as particles. In a statistical sense, elastic scattering of light by numerous particles much smaller than the wavelength of the light is a process known as Rayleigh scattering while the similar process for scattering by particles that are similar or larger in wavelength is known as Mie scattering with the Tyndall effect being a commonly observed result. A small proportion of light scattering from atoms or molecules may undergo Raman scattering, wherein the frequency changes due to excitation of the atoms and molecules. Brillouin scattering occurs when the frequency of light changes due to local changes with time and movements of a dense material.
Dispersion occurs when different frequencies of light have different phase velocities, due either to material properties (material dispersion) or to the geometry of an optical waveguide (waveguide dispersion). The most familiar form of dispersion is a decrease in index of refraction with increasing wavelength, which is seen in most transparent materials. This is called "normal dispersion". It occurs in all dielectric materials, in wavelength ranges where the material does not absorb light.] In wavelength ranges where a medium has significant absorption, the index of refraction can increase with wavelength. This is called "anomalous dispersion".
The separation of colours by a prism is an example of normal dispersion. At the surfaces of the prism, Snell's law predicts that light incident at an angle θ to the normal will be refracted at an angle arcsin(sin (θ) / n). Thus, blue light, with its higher refractive index, is bent more strongly than red light, resulting in the well-known rainbow pattern.

Dispersion: two sinusoids propagating at different speeds make a moving interference pattern. The red dot moves with the phase velocity, and the green dots propagate with the group velocity. In this case, the phase velocity is twice the group velocity. The red dot overtakes two green dots, when moving from the left to the right of the figure. In effect, the individual waves (which travel with the phase velocity) escape from the wave packet (which travels with the group velocity).
Material dispersion is often characterised by the Abbe number, which gives a simple measure of dispersion based on the index of refraction at three specific wavelengths. Waveguide dispersion is dependent on the propagation constant. Both kinds of dispersion cause changes in the group characteristics of the wave, the features of the wave packet that change with the same frequency as the amplitude of the electromagnetic wave. "Group velocity dispersion" manifests as a spreading-out of the signal "envelope" of the radiation and can be quantified with a group dispersion delay parameter:
where is the group velocity] For a uniform medium, the group velocity is
where n is the index of refraction and c is the speed of light in a vacuum. This gives a simpler form for the dispersion delay parameter:
If D is less than zero, the medium is said to have positive dispersion or normal dispersion. If D is greater than zero, the medium has negative dispersion. If a light pulse is propagated through a normally dispersive medium, the result is the higher frequency components slow down more than the lower frequency components. The pulse therefore becomes positively chirped, or up-chirped, increasing in frequency with time. This causes the spectrum coming out of a prism to appear with red light the least refracted and blue/violet light the most refracted. Conversely, if a pulse travels through an anomalously (negatively) dispersive medium, high frequency components travel faster than the lower ones, and the pulse becomes negatively chirped, or down-chirped, decreasing in frequency with time.
The result of group velocity dispersion, whether negative or positive, is ultimately temporal spreading of the pulse. This makes dispersion management extremely important in optical communications systems based on optical fibres, since if dispersion is too high, a group of pulses representing information will each spread in time and merge, making it impossible to extract the signal.

Polarization

Polarization is a general property of waves that describes the orientation of their oscillations. For transverse waves such as many electromagnetic waves, it describes the orientation of the oscillations in the plane perpendicular to the wave's direction of travel. The oscillations may be oriented in a single direction (linear polarization), or the oscillation direction may rotate as the wave travels (circular or elliptical polarization). Circularly polarised waves can rotate rightward or leftward in the direction of travel, and which of those two rotations is present in a wave is called the wave's chirality.
The typical way to consider polarization is to keep track of the orientation of the electric field vector as the electromagnetic wave propagates. The electric field vector of a plane wave may be arbitrarily divided into two perpendicular components labeled x and y (with z indicating the direction of travel). The shape traced out in the x-y plane by the electric field vector is a Lissajous figure that describes the polarization state] The following figures show some examples of the evolution of the electric field vector (blue), with time (the vertical axes), at a particular point in space, along with its x and y components (red/left and green/right), and the path traced by the vector in the plane (purple): The same evolution would occur when looking at the electric field at a particular time while evolving the point in space, along the direction opposite to propagation.
Linear polarization diagram
Linear
Circular polarization diagram
Circular
Elliptical polarization diagram
Elliptical polarization
In the leftmost figure above, the x and y components of the light wave are in phase. In this case, the ratio of their strengths is constant, so the direction of the electric vector (the vector sum of these two components) is constant. Since the tip of the vector traces out a single line in the plane, this special case is called linear polarization. The direction of this line depends on the relative amplitudes of the two components.
In the middle figure, the two orthogonal components have the same amplitudes and are 90° out of phase. In this case, one component is zero when the other component is at maximum or minimum amplitude. There are two possible phase relationships that satisfy this requirement: the x component can be 90° ahead of the y component or it can be 90° behind the y component. In this special case, the electric vector traces out a circle in the plane, so this polarization is called circular polarization. The rotation direction in the circle depends on which of the two phase relationships exists and corresponds to right-hand circular polarization and left-hand circular polarization.
In all other cases, where the two components either do not have the same amplitudes and/or their phase difference is neither zero nor a multiple of 90°, the polarization is called elliptical polarization because the electric vector traces out an ellipse in the plane (the polarization ellipse). This is shown in the above figure on the right. Detailed mathematics of polarization is done using Jones calculus and is characterised by the Stokes parameters.
Changing polarization
Media that have different indexes of refraction for different polarization modes are called birefringent. Well known manifestations of this effect appear in optical wave plates/retarders (linear modes) and in Faraday rotation/optical rotation (circular modes). If the path length in the birefringent medium is sufficient, plane waves will exit the material with a significantly different propagation direction, due to refraction. For example, this is the case with macroscopic crystals of calcite, which present the viewer with two offset, orthogonally polarised images of whatever is viewed through them. It was this effect that provided the first discovery of polarization, by Erasmus Bartholinus in 1669. In addition, the phase shift, and thus the change in polarization state, is usually frequency dependent, which, in combination with dichroism, often gives rise to bright colours and rainbow-like effects. In mineralogy, such properties, known as pleochroism, are frequently exploited for the purpose of identifying minerals using polarization microscopes. Additionally, many plastics that are not normally birefringent will become so when subject to mechanical stress, a phenomenon which is the basis of photoelasticity] Non-birefringent methods, to rotate the linear polarization of light beams, include the use of prismatic polarization rotators which use total internal reflection in a prism set designed for efficient collinear transmission.
A polariser changing the orientation of linearly polarised light.
In this picture, θ1θ0 = θi.
Media that reduce the amplitude of certain polarization modes are called dichroic, with devices that block nearly all of the radiation in one mode known as polarizing filters or simply "polarisers". Malus' law, which is named after Étienne-Louis Malus, says that when a perfect polariser is placed in a linear polarised beam of light, the intensity, I, of the light that passes through is given by
where
I0 is the initial intensity,
and θi is the angle between the light's initial polarization direction and the axis of the polariser.
A beam of unpolarised light can be thought of as containing a uniform mixture of linear polarizations at all possible angles. Since the average value of is 1/2, the transmission coefficient becomes
In practice, some light is lost in the polariser and the actual transmission of unpolarised light will be somewhat lower than this, around 38% for Polaroid-type polarisers but considerably higher (>49.9%) for some birefringent prism types.
In addition to birefringence and dichroism in extended media, polarization effects can also occur at the (reflective) interface between two materials of different refractive index. These effects are treated by the Fresnel equations. Part of the wave is transmitted and part is reflected, with the ratio depending on angle of incidence and the angle of refraction. In this way, physical optics recovers Brewster's angle ] When light reflects from a thin film on a surface, interference between the reflections from the film's surfaces can produce polarization in the reflected and transmitted light.
Natural light
The effects of a polarising filter on the sky in a photograph. Left picture is taken without polariser. For the right picture, filter was adjusted to eliminate certain polarizations of the scattered blue light from the sky.
Most sources of electromagnetic radiation contain a large number of atoms or molecules that emit light. The orientation of the electric fields produced by these emitters may not be correlated, in which case the light is said to be unpolarised. If there is partial correlation between the emitters, the light is partially polarised. If the polarization is consistent across the spectrum of the source, partially polarised light can be described as a superposition of a completely unpolarised component, and a completely polarised one. One may then describe the light in terms of the degree of polarization, and the parameters of the polarization ellipse.
Light reflected by shiny transparent materials is partly or fully polarised, except when the light is normal (perpendicular) to the surface. It was this effect that allowed the mathematician Étienne-Louis Malus to make the measurements that allowed for his development of the first mathematical models for polarised light. Polarization occurs when light is scattered in the atmosphere. The scattered light produces the brightness and colour in clear skies. This partial polarization of scattered light can be taken advantage of using polarizing filters to darken the sky in photographs. Optical polarization is principally of importance in chemistry due to circular dichroism and optical rotation ("circular birefringence") exhibited by optically active (chiral) molecules .

Modern optics

Modern optics encompasses the areas of optical science and engineering that became popular in the 20th century. These areas of optical science typically relate to the electromagnetic or quantum properties of light but do include other topics. A major subfield of modern optics, quantum optics, deals with specifically quantum mechanical properties of light. Quantum optics is not just theoretical; some modern devices, such as lasers, have principles of operation that depend on quantum mechanics. Light detectors, such as photomultipliers and channeltrons, respond to individual photons. Electronic image sensors, such as CCDs, exhibit shot noise corresponding to the statistics of individual photon events. Light-emitting diodes and photovoltaic cells, too, cannot be understood without quantum mechanics. In the study of these devices, quantum optics often overlaps with quantum electronics.
Specialty areas of optics research include the study of how light interacts with specific materials as in crystal optics and metamaterials. Other research focuses on the phenomenology of electromagnetic waves as in singular optics, non-imaging optics, non-linear optics, statistical optics, and radiometry. Additionally, computer engineers have taken an interest in integrated optics, machine vision, and photonic computing as possible components of the "next generation" of computers.
Today, the pure science of optics is called optical science or optical physics to distinguish it from applied optical sciences, which are referred to as optical engineering. Prominent subfields of optical engineering include illumination engineering, photonics, and optoelectronics with practical applications like lens design, fabrication and testing of optical components, and image processing. Some of these fields overlap, with nebulous boundaries between the subjects terms that mean slightly different things in different parts of the world and in different areas of industry. A professional community of researchers in nonlinear optics has developed in the last several decades due to advances in laser technology.

Lasers

Experiments such as this one with high-power lasers are part of the modern optics research.
A laser is a device that emits light (electromagnetic radiation) through a process called stimulated emission. The term laser is an acronym for Light Amplification by Stimulated Emission of Radiation] Laser light is usually spatially coherent, which means that the light either is emitted in a narrow, low-divergence beam, or can be converted into one with the help of optical components such as lenses. Because the microwave equivalent of the laser, the maser, was developed first, devices that emit microwave and radio frequencies are usually called masers.
VLT’s laser guided star.
The first working laser was demonstrated on 16 May 1960 by Theodore Maiman at Hughes Research Laboratories. When first invented, they were called "a solution looking for a problem".] Since then, lasers have become a multibillion-dollar industry, finding utility in thousands of highly varied applications. The first application of lasers visible in the daily lives of the general population was the supermarket barcode scanner, introduced in 1974. The laserdisc player, introduced in 1978, was the first successful consumer product to include a laser, but the compact disc player was the first laser-equipped device to become truly common in consumers' homes, beginning in 1982. These optical storage devices use a semiconductor laser less than a millimetre wide to scan the surface of the disc for data retrieval. Fibre-optic communication relies on lasers to transmit large amounts of information at the speed of light. Other common applications of lasers include laser printers and laser pointers. Lasers are used in medicine in areas such as bloodless surgery, laser eye surgery, and laser capture microdissection and in military applications such as missile defence systems, electro-optical countermeasures (EOCM), and lidar. Lasers are also used in holograms, bubblegrams, laser light shows, and laser hair removal.

Kapitsa–Dirac effect

The Kapitsa–Dirac effect causes beams of particles to diffract as the result of meeting a standing wave of light. Light can be used to position matter using various phenomena (see optical tweezers).

Applications

Optics is part of everyday life. The ubiquity of visual systems in biology indicates the central role optics plays as the science of one of the five senses. Many people benefit from eyeglasses or contact lenses, and optics are integral to the functioning of many consumer goods including cameras. Rainbows and mirages are examples of optical phenomena. Optical communication provides the backbone for both the Internet and modern telephony.

Human eye

Model of a human eye. Features mentioned in this article are 3. ciliary muscle, 6. pupil, 8. cornea, 10. lens cortex, 22. optic nerve, 26. fovea, 30. retina
The human eye functions by focusing light onto a layer of photoreceptor cells called the retina, which forms the inner lining of the back of the eye. The focusing is accomplished by a series of transparent media. Light entering the eye passes first through the cornea, which provides much of the eye's optical power. The light then continues through the fluid just behind the cornea—the anterior chamber, then passes through the pupil. The light then passes through the lens, which focuses the light further and allows adjustment of focus. The light then passes through the main body of fluid in the eye—the vitreous humour, and reaches the retina. The cells in the retina line the back of the eye, except for where the optic nerve exits; this results in a blind spot.
There are two types of photoreceptor cells, rods and cones, which are sensitive to different aspects of light.Rod cells are sensitive to the intensity of light over a wide frequency range, thus are responsible for black-and-white vision. Rod cells are not present on the fovea, the area of the retina responsible for central vision, and are not as responsive as cone cells to spatial and temporal changes in light. There are, however, twenty times more rod cells than cone cells in the retina because the rod cells are present across a wider area. Because of their wider distribution, rods are responsible for peripheral vision.
In contrast, cone cells are less sensitive to the overall intensity of light, but come in three varieties that are sensitive to different frequency-ranges and thus are used in the perception of colour and photopic vision. Cone cells are highly concentrated in the fovea and have a high visual acuity meaning that they are better at spatial resolution than rod cells. Since cone cells are not as sensitive to dim light as rod cells, most night vision is limited to rod cells. Likewise, since cone cells are in the fovea, central vision (including the vision needed to do most reading, fine detail work such as sewing, or careful examination of objects) is done by cone cells.
Ciliary muscles around the lens allow the eye's focus to be adjusted. This process is known as accommodation. The near point and far point define the nearest and farthest distances from the eye at which an object can be brought into sharp focus. For a person with normal vision, the far point is located at infinity. The near point's location depends on how much the muscles can increase the curvature of the lens, and how inflexible the lens has become with age. Optometrists, ophthalmologists, and opticians usually consider an appropriate near point to be closer than normal reading distance—approximately 25 cm.
Defects in vision can be explained using optical principles. As people age, the lens becomes less flexible and the near point recedes from the eye, a condition known as presbyopia. Similarly, people suffering from hyperopia cannot decrease the focal length of their lens enough to allow for nearby objects to be imaged on their retina. Conversely, people who cannot increase the focal length of their lens enough to allow for distant objects to be imaged on the retina suffer from myopia and have a far point that is considerably closer than infinity. A condition known as astigmatism results when the cornea is not spherical but instead is more curved in one direction. This causes horizontally extended objects to be focused on different parts of the retina than vertically extended objects, and results in distorted images.
All of these conditions can be corrected using corrective lenses. For presbyopia and hyperopia, a converging lens provides the extra curvature necessary to bring the near point closer to the eye while for myopia a diverging lens provides the curvature necessary to send the far point to infinity. Astigmatism is corrected with a cylindrical surface lens that curves more strongly in one direction than in another, compensating for the non-uniformity of the cornea.
The optical power of corrective lenses is measured in diopters, a value equal to the reciprocal of the focal length measured in metres; with a positive focal length corresponding to a converging lens and a negative focal length corresponding to a diverging lens. For lenses that correct for astigmatism as well, three numbers are given: one for the spherical power, one for the cylindrical power, and one for the angle of orientation of the astigmatism.

Visual effects

The Ponzo Illusion relies on the fact that parallel lines appear to converge as they approach infinity.
Optical illusions (also called visual illusions) are characterized by visually perceived images that differ from objective reality. The information gathered by the eye is processed in the brain to give a percept that differs from the object being imaged. Optical illusions can be the result of a variety of phenomena including physical effects that create images that are different from the objects that make them, the physiological effects on the eyes and brain of excessive stimulation (e.g. brightness, tilt, colour, movement), and cognitive illusions where the eye and brain make unconscious inferences.[82]
Cognitive illusions include some which result from the unconscious misapplication of certain optical principles. For example, the Ames room, Hering, Müller-Lyer, Orbison, Ponzo, Sander, and Wundt illusions all rely on the suggestion of the appearance of distance by using converging and diverging lines, in the same way that parallel light rays (or indeed any set of parallel lines) appear to converge at a vanishing point at infinity in two-dimensionally rendered images with artistic perspective.[83] This suggestion is also responsible for the famous moon illusion where the moon, despite having essentially the same angular size, appears much larger near the horizon than it does at zenith.[84] This illusion so confounded Ptolemy that he incorrectly attributed it to atmospheric refraction when he described it in his treatise, Optics.
Another type of optical illusion exploits broken patterns to trick the mind into perceiving symmetries or asymmetries that are not present. Examples include the café wall, Ehrenstein, Fraser spiral, Poggendorff, and Zöllner illusions. Related, but not strictly illusions, are patterns that occur due to the superimposition of periodic structures. For example, transparent tissues with a grid structure produce shapes known as moiré patterns, while the superimposition of periodic transparent patterns comprising parallel opaque lines or curves produces line moiré patterns.

Optical instruments

Illustrations of various optical instruments from the 1728 Cyclopaedia
Single lenses have a variety of applications including photographic lenses, corrective lenses, and magnifying glasses while single mirrors are used in parabolic reflectors and rear-view mirrors. Combining a number of mirrors, prisms, and lenses produces compound optical instruments which have practical uses. For example, a periscope is simply two plane mirrors aligned to allow for viewing around obstructions. The most famous compound optical instruments in science are the microscope and the telescope which were both invented by the Dutch in the late 16th century.
Microscopes were first developed with just two lenses: an objective lens and an eyepiece. The objective lens is essentially a magnifying glass and was designed with a very small focal length while the eyepiece generally has a longer focal length. This has the effect of producing magnified images of close objects. Generally, an additional source of illumination is used since magnified images are dimmer due to the conservation of energy and the spreading of light rays over a larger surface area. Modern microscopes, known as compound microscopes have many lenses in them (typically four) to optimize the functionality and enhance image stability. A slightly different variety of microscope, the comparison microscope, looks at side-by-side images to produce a stereoscopic binocular view that appears three dimensional when used by humans.
The first telescopes, called refracting telescopes were also developed with a single objective and eyepiece lens. In contrast to the microscope, the objective lens of the telescope was designed with a large focal length to avoid optical aberrations. The objective focuses an image of a distant object at its focal point which is adjusted to be at the focal point of an eyepiece of a much smaller focal length. The main goal of a telescope is not necessarily magnification, but rather collection of light which is determined by the physical size of the objective lens. Thus, telescopes are normally indicated by the diameters of their objectives rather than by the magnification which can be changed by switching eyepieces. Because the magnification of a telescope is equal to the focal length of the objective divided by the focal length of the eyepiece, smaller focal-length eyepieces cause greater magnification.
Since crafting large lenses is much more difficult than crafting large mirrors, most modern telescopes are reflecting telescopes, that is, telescopes that use a primary mirror rather than an objective lens. The same general optical considerations apply to reflecting telescopes that applied to refracting telescopes, namely, the larger the primary mirror, the more light collected, and the magnification is still equal to the focal length of the primary mirror divided by the focal length of the eyepiece. Professional telescopes generally do not have eyepieces and instead place an instrument (often a charge-coupled device) at the focal point instead.

Photography

Photograph taken with aperture f/32
Photograph taken with aperture f/5
The optics of photography involves both lenses and the medium in which the electromagnetic radiation is recorded, whether it be a plate, film, or charge-coupled device. Photographers must consider the reciprocity of the camera and the shot which is summarized by the relation
Exposure ∝ ApertureArea × ExposureTime × SceneLuminance
In other words, the smaller the aperture (giving greater depth of focus), the less light coming in, so the length of time has to be increased (leading to possible blurriness if motion occurs). An example of the use of the law of reciprocity is the Sunny 16 rule which gives a rough estimate for the settings needed to estimate the proper exposure in daylight.
A camera's aperture is measured by a unitless number called the f-number or f-stop, f/#, often notated as , and given by
where is the focal length, and is the diameter of the entrance pupil. By convention, "f/#" is treated as a single symbol, and specific values of f/# are written by replacing the number sign with the value. The two ways to increase the f-stop are to either decrease the diameter of the entrance pupil or change to a longer focal length (in the case of a zoom lens, this can be done by simply adjusting the lens). Higher f-numbers also have a larger depth of field due to the lens approaching the limit of a pinhole camera which is able to focus all images perfectly, regardless of distance, but requires very long exposure times.
The field of view that the lens will provide changes with the focal length of the lens. There are three basic classifications based on the relationship to the diagonal size of the film or sensor size of the camera to the focal length of the lens:
  • Normal lens: angle of view of about 50° (called normal because this angle considered roughly equivalent to human vision ) and a focal length approximately equal to the diagonal of the film or sensor
  • Wide-angle lens: angle of view wider than 60° and focal length shorter than a normal lens.
  • Long focus lens: angle of view narrower than a normal lens. This is any lens with a focal length longer than the diagonal measure of the film or sensor. The most common type of long focus lens is the telephoto lens, a design that uses a special telephoto group to be physically shorter than its focal length.
Modern zoom lenses may have some or all of these attributes.
The absolute value for the exposure time required depends on how sensitive to light the medium being used is (measured by the film speed, or, for digital media, by the quantum efficiency).[96] Early photography used media that had very low light sensitivity, and so exposure times had to be long even for very bright shots. As technology has improved, so has the sensitivity through film cameras and digital cameras.
Other results from physical and geometrical optics apply to camera optics. For example, the maximum resolution capability of a particular camera set-up is determined by the diffraction limit associated with the pupil size and given, roughly, by the Rayleigh criterion.

Atmospheric optics

A colourful sky is often due to scattering of light off particulates and pollution, as in this photograph of a sunset during the October 2007 California wildfires.
The unique optical properties of the atmosphere cause a wide range of spectacular optical phenomena. The blue colour of the sky is a direct result of Rayleigh scattering which redirects higher frequency (blue) sunlight back into the field of view of the observer. Because blue light is scattered more easily than red light, the sun takes on a reddish hue when it is observed through a thick atmosphere, as during a sunrise or sunset. Additional particulate matter in the sky can scatter different colours at different angles creating colourful glowing skies at dusk and dawn. Scattering off of ice crystals and other particles in the atmosphere are responsible for halos, afterglows, coronas, rays of sunlight, and sun dogs. The variation in these kinds of phenomena is due to different particle sizes and geometries.
Mirages are optical phenomena in which light rays are bent due to thermal variations in the refraction index of air, producing displaced or heavily distorted images of distant objects. Other dramatic optical phenomena associated with this include the Novaya Zemlya effect where the sun appears to rise earlier than predicted with a distorted shape. A spectacular form of refraction occurs with a temperature inversion called the Fata Morgana where objects on the horizon or even beyond the horizon, such as islands, cliffs, ships or icebergs, appear elongated and elevated, like "fairy tale castles".
Rainbows are the result of a combination of internal reflection and dispersive refraction of light in raindrops. A single reflection off the backs of an array of raindrops produces a rainbow with an angular size on the sky that ranges from 40° to 42° with red on the outside. Double rainbows are produced by two internal reflections with angular size of 50.5° to 54° with violet on the outside. Because rainbows are seen with the sun 180° away from the centre of the rainbow, rainbows are more prominent the closer the sun is to the horizon

   

________________________________________

                    Optical computing

Optical or photonic computing uses photons produced by lasers or diodes for computation. For decades, photons have promised to allow a higher bandwidth than the electrons used in conventional computers.
Most research projects focus on replacing current computer components with optical equivalents, resulting in an optical digital computer system processing binary data. This approach appears to offer the best short-term prospects for commercial optical computing, since optical components could be integrated into traditional computers to produce an optical-electronic hybrid. However, optoelectronic devices lose 30% of their energy converting electronic energy into photons and back; this conversion also slows the transmission of messages. All-optical computers eliminate the need for optical-electrical-optical (OEO) conversions, thus lessening the need for electrical power.
Application-specific devices, such as synthetic aperture radar (SAR) and optical correlators, have been designed to use the principles of optical computing. Correlators can be used, for example, to detect and track objects, and to classify serial time-domain optical data

Optical components for binary digital computer

The fundamental building block of modern electronic computers is the transistor. To replace electronic components with optical ones, an equivalent optical transistor is required. This is achieved using materials with a non-linear refractive index. In particular, materials exist where the intensity of incoming light affects the intensity of the light transmitted through the material in a similar manner to the current response of a bipolar transistor. Such an optical transistor  can be used to create optical logic gates, which in turn are assembled into the higher level components of the computer's CPU. These will be nonlinear optical crystals used to manipulate light beams into controlling other light beams.
Like any computing system, an Optical computing system needs three things to function well:
  1. optical processor
  2. optical data transfer, e.g. Fiber optic cable
  3. optical storage, e.g. CD/DVD/bluray, etc.
Substituting electrical components will need data format conversion from photons to electrons, which will make the system slower.

Controversy

There are disagreements between researchers about the future capabilities of optical computers; whether or not they may be able to compete with semiconductor-based electronic computers in terms of speed, power consumption, cost, and size is an open question. Critics note that real-world logic systems require "logic-level restoration, cascadability, fan-out and input–output isolation", all of which are currently provided by electronic transistors at low cost, low power, and high speed. For optical logic to be competitive beyond a few niche applications, major breakthroughs in non-linear optical device technology would be required, or perhaps a change in the nature of computing itself.

Misconceptions, challenges, and prospects

A significant challenge to optical computing is that computation is a nonlinear process in which multiple signals must interact. Light, which is an electromagnetic wave, can only interact with another electromagnetic wave in the presence of electrons in a material, and the strength of this interaction is much weaker for electromagnetic waves, such as light, than for the electronic signals in a conventional computer. This may result in the processing elements for an optical computer requiring more power and larger dimensions than those for a conventional electronic computer using transistors.[citation needed]
A further misconception is that since light can travel much faster than the drift velocity of electrons, and at frequencies measured in THz, optical transistors should be capable of extremely high frequencies. However, any electromagnetic wave must obey the transform limit, and therefore the rate at which an optical transistor can respond to a signal is still limited by its spectral bandwidth. However, in fiber optic communications, practical limits such as dispersion often constrain channels to bandwidths of 10s of GHz, only slightly better than many silicon transistors. Obtaining dramatically faster operation than electronic transistors would therefore require practical methods of transmitting ultrashort pulses down highly dispersive waveguides.

Photonic logic

Realization of a photonic controlled-NOT gate for use in quantum computing
Photonic logic is the use of photons (light) in logic gates (NOT, AND, OR, NAND, NOR, XOR, XNOR). Switching is obtained using nonlinear optical effects when two or more signals are combined.
Resonators are especially useful in photonic logic, since they allow a build-up of energy from constructive interference, thus enhancing optical nonlinear effects.
Other approaches that have been investigated include photonic logic at a molecular level, using photoluminescent chemicals. In a demonstration, Witlicki et al. performed logical operations using molecules and SERS.

Unconventional approaches

Time delays optical computing

The basic idea is to delay light (or any other signal) in order to perform useful computations. Of interest would be to solve NP-complete problems as those are difficult problems for the conventional computers.
There are 2 basic properties of light that are actually used in this approach:
  • The light can be delayed by passing it through an optical fiber of a certain length.
  • The light can be split into multiple (sub)rays. This property is also essential because we can evaluate multiple solutions in the same time.
When solving a problem with time-delays the following steps must be followed:
  • The first step is to create a graph-like structure made from optical cables and splitters. Each graph has a start node and a destination node.
  • The light enters through the start node and traverses the graph until it reaches the destination. It is delayed when passing through arcs and divided inside nodes.
  • The light is marked when passing through an arc or through an node so that we can easily identify that fact at the destination node.
  • At the destination node we will wait for a signal (fluctuation in the intensity of the signal) which arrives at a particular moment(s) in time. If there is no signal arriving at that moment, it means that we have no solution for our problem. Otherwise the problem has a solution. Fluctuations can be read with a photodetector and an oscilloscope.
The first problem attacked in this way was the Hamiltonian path problem.
The simplest one is the subset sum problem. An optical device solving an instance with 4 numbers {a1, a2, a3, a4} is depicted below:
Optical device for solving the Subset sum problem
The light will enter in Start node. It will be divided into 2 (sub)rays of smaller intensity. These 2 rays will arrive into the second node at moments a1 and 0. Each of them will be divided into 2 subrays which will arrive in the 3rd node at moments 0, a1, a2 and a1 + a2. These represents the all subsets of the set {a1, a2}. We expect fluctuations in the intensity of the signal at no more than 4 different moments. In the destination node we expect fluctuations at no more than 16 different moments (which are all the subsets of the given. If we have a fluctuation in the target moment B, it means that we have a solution of the problem, otherwise there is no subset whose sum of elements equals B. For the practical implementation we cannot have zero-length cables, thus all cables are increased with a small (fixed for all) value k. In this case the solution is expected at moment B+n*k.

Wavelength-based computing

Wavelength-based computing can be used to solve the 3-SAT problem with n variables, m clause and with no more than 3 variables per clause. Each wavelength, contained in a light ray, is considered as possible value-assignments to n variables. The optical device contains prisms and mirrors are used to discriminate proper wavelengths which satisfy the formula.

Computing by xeroxing on transparencies

This approach uses a Xerox machine and transparent sheets for performing computations. k-SAT problem with n variables, m clauses and at most k variables per clause has been solved in 3 steps:
  • Firstly all 2^n possible assignments of n variables have been generated by performing n xerox copies.
  • Using at most 2k copies of the truth table, each clause is evaluated at every row of the truth table simultaneously.
  • The solution is obtained by making a single copy operation of the overlapped transparencies of all m clauses.

Masking optical beams

The travelling salesman problem has been solved in by using an optical approach. All possible TSP paths have been generated and stored in a binary matrix which was multiplied with another gray-scale vector containing the distances between cities. The multiplication is performed optically by using an optical correlator.

Optical Fourier co-processors

Many computations, particularly in scientific applications, require frequent use of the 2D discrete Fourier transform (DFT) – for example in solving differential equations describing propagation of waves or transfer of heat. Though modern GPU technologies typically enable high-speed computation of large 2D DFTs, techniques have been developed that can perform DFTs optically by utilising the natural Fourier transforming property of lenses. The input is encoded using a liquid crystal spatial light modulator and the result is measured using a conventional CMOS or CCD image sensor. Such optical architectures can offer superior scaling of computational complexity due to the inherently highly interconnected nature of optical propagation, and have been used to solve 2D heat equations.

Ising machines

Physical computers whose design was inspired by the theoretical Ising model are called Ising machines.
Yoshihisa Yamamoto pioneered building Ising machines using photons. Initially Yamamoto and his colleagues built an Ising machine using lasers, mirrors, and other optical components commonly found on an optical table.[18][19]
Later a team at Hewlett Packard Labs including Dave Kielpinski developed photonic chip design tools and used them to build an Ising machine on a single chip, integrating 1,052 optical components on that single chip.


________________________________________________________________________________


  

                      Optical transistor

An optical transistor, also known as an optical switch or a light valve, is a device that switches or amplifies optical signals. Light occurring on an optical transistor’s input changes the intensity of light emitted from the transistor’s output while output power is supplied by an additional optical source. Since the input signal intensity may be weaker than that of the source, an optical transistor amplifies the optical signal. The device is the optical analog of the electronic transistor that forms the basis of modern electronic devices. Optical transistors provide a means to control light using only light and has applications in optical computing and fiber-optic communication networks. Such technology has the potential to exceed the speed of electronics, while saving more power.
Since photons inherently do not interact with each other, an optical transistor must employ an operating medium to mediate interactions. This is done without converting optical to electronic signals as an intermediate step. Implementations using a variety of operating mediums have been proposed and experimentally demonstrated. However, their ability to compete with modern electronics is currently limited.

Applications

Optical transistors could be used to improve the performance of fiber-optic communication networks. Although fiber-optic cables are used to transfer data, tasks such as signal routing are done electronically. This requires optical-electronic-optical conversion, which form bottlenecks. In principle, all-optical digital signal processing and routing is achievable using optical transistors arranged into photonic integrated circuits [1]. The same devices could be used to create new types of optical amplifiers to compensate for signal attenuation along transmission lines.
A more elaborate application of optical transistors is the development of an optical digital computer in which components process photons rather than electrons. Further, optical transistors that operate using single photons could form an integral part of quantum information processing where they can be used to selectively address individual units of quantum information, known as qubits.

Comparison with electronics

The most commonly argued case for optical logic is that optical transistor switching times can be much faster than in conventional electronic transistors. This is due to the fact that the speed of light in an optical medium is typically much faster than the drift velocity of electrons in semiconductors.
Optical transistors can be directly linked to fiber-optic cables whereas electronics requires coupling via photodetectors and LEDs or lasers. The more natural integration of all-optical signal processors with fiber-optics would reduce the complexity and delay in the routing and other processing of signals in optical communication networks.
It remains questionable whether optical processing can reduce the energy required to switch a single transistor to be less than that for electronic transistors. To realistically compete, transistors require a few tens of photons per operation. It is clear, however, that this is achievable in proposed single-photon transistors for quantum information processing.
Perhaps the most significant advantage of optical over electronic logic is reduced power consumption. This comes from the absence of capacitance in the connections between individual logic gates. In electronics, the transmission line needs to be charged to the signal voltage. The capacitance of a transmission line is proportional to its length and it exceeds the capacitance of the transistors in a logic gate when its length is equal to that of a single gate. The charging of transmission lines is one of the main energy losses in electronic logic. This loss is avoided in optical communication where only enough energy to switch an optical transistor at the receiving end must be transmitted down a line. This fact has played a major role in the uptake of fiber optics for long distance communication but is yet to be exploited at the microprocessor level.
Besides the potential advantages of higher speed, lower power consumption and high compatibility with optical communication systems, optical transistors must satisfy a set of benchmarks before they can compete with electronics. No single design has yet satisfied all these criteria whilst outperforming speed and power consumption of state of the art electronics.
The criteria include:
  • Fan-out - Transistor output must be in the correct form and of sufficient power to operate the inputs of at least two transistors. This implies that the input and output wavelengths, beam shapes and pulse shapes must be compatible.
  • Logic level restoration - The signal needs to be ‘cleaned’ by each transistor. Noise and degradations in signal quality must be removed so that they do not propagate through the system and accumulate to produce errors.
  • Logic level independent of loss - In optical communication, the signal intensity decreases over distance due to absorption of light in the fiber optic cable. Therefore, a simple intensity threshold cannot distinguish between on and off signals for arbitrary length interconnects. The system must encode zeros and ones at different frequencies, use differential signaling where the ratio or difference in two different powers carries the logic signal to avoid errors.

Implementations

Several schemes have been proposed to implement all-optical transistors. In many cases, a proof of concept has been experimentally demonstrated. Among the designs are those based on:
  • electromagnetically induced transparency
    • in an optical cavity or microresonator, where the transmission is controlled by a weaker flux of gate photons
    • in free space, i.e., without a resonator, by addressing strongly interacting Rydberg states
  • a system of indirect excitons (composed of bound pairs of electrons and holes in double quantum wells with a static dipole moment). Indirect excitons, which are created by light and decay to emit light, strongly interact due to their dipole alignment.
  • a system of microcavity polaritons (exciton-polaritons inside an optical microcavity) where, similar to exciton-based optical transistors, polaritons facilitate effective interactions between photon
  • photonic crystal cavities with an active Raman gain medium
  • cavity switch modulates cavity properties in time domain for quantum information applications 
  • nanowire-based cavities employing polaritonic interactions for optical switching
  • silicon microrings placed in the path of an optical signal. Gate photons heat the silicon microring causing a shift in the optical resonant frequency, leading to a change in transparency at a given frequency of the optical supply.
  • a dual-mirror optical cavity that holds around 20,000 cesium atoms trapped by means of optical tweezers and laser-cooled to a few microkelvin. The cesium ensemble did not interact with light and was thus transparent. The length of a round trip between the cavity mirrors equaled an integer multiple of the wavelength of the incident light source, allowing the cavity to transmit the source light. Photons from the gate light field entered the cavity from the side, where each photon interacted with an additional "control" light field, changing a single atom's state to be resonant with the cavity optical field, which changing the field's resonance wavelength and blocking transmission of the source field, thereby "switching" the "device". While the changed atom remains unidentified, quantum interference allows the gate photon to be retrieved from the cesium. A single gate photon could redirect a source field containing up to two photons before the retrieval of the gate photon was impeded, above the critical threshold for a positive gain


_________________________________________________________________________________


    

                         Optical switch

An optical switch is a device that selectively switches optical signals on or off or from one channel to another. The former is known as an optical (time-domain) switch or an optical modulator, while the latter is called an optical space switch or an optical router. Since the switching can be temporal or spatial, such switches are analogous to one-way or two-way switches in electrical circuits. In general, optical modulators and routers can be made from each other.

Terminology

The word applies on several levels. In commercial terms (such as "the telecom optical switch market size") it refers to any piece of circuit switching equipment between fibers. The majority of installed systems in this category actually use electronic switching between fiber transponders. Systems that perform this function by routing light beams are often referred to as "photonic" switches, independent of how the light itself is switched. Away from telecom, an optical switch is the unit that actually switches light between fibers, and a photonic switch is one that does this by exploiting nonlinear material properties, such as semiconductor-based materials, to steer light (i.e., to switch wavelengths, intensities, or directions) . Hence a certain portion of the optical switch market is made up of photonic switches. These will contain within them an optical switch, which will, in some cases, be a photonic switch.

Operation

An optical switch may operate by mechanical means, such as physically shifting an optical fiber to drive one or more alternative fibers, or by electro-optic effects, magneto-optic effects, or other methods. Slow optical switches, such as those using moving fibers, may be used for alternate routing of an optical switch transmission path, such as routing around a fault. Fast optical switches, such as those using electro-optic or magneto-optic effects, may be used to perform logic operations; also included in this category are semiconductor optical amplifiers, which are optoelectronic devices that can be used as optical switches and be integrated with discrete or integrated microelectronic circuits.

Functionality

The functionality of any switch can be described in terms of the connections it can establish. As stated in Telcordia GR-1073, a connection is the association between two ports on a switch and is indicated as a pair of port identifiers (i, j ), where i and j are two ports between which the connection is established. A connection identifies the transmission path between two ports. An optical signal can be applied to either one of the connected ports. However, the nature of the signal emerging at the other port depends on the optical switch and the state of the connection. A connection can be in the on state or the off state. A connection is said to be in the on state if an optical signal applied to one port emerges at the other port with essentially zero loss in optical energy. A connection is said to be in the off state if essentially zero optical energy emerges at the other port.
Connections established in optical switches can be unidirectional or bidirectional. A unidirectional connection only allows optical signal transmission in one direction between the connected ports. A bidirectional connection allows optical signal transmission in both directions over the connection. Connections in passive and transparent optical switches are bidirectional, i.e., if a connection (i, j ) is set up, optical transmission is possible from i to j and from j to i.
A device is optically “transparent” if the optical signal launched at the input remains optical throughout its transmission path in the device and appears as an optical signal at the output. Optically transparent devices operate over a range of wavelengths called the passband.
A passive optical switch does not have optical gain elements. An active optical switch has optical gain elements. An all-optical switch is a transparent optical switch in which the actuating signal is also optical. Thus, in an all-optical switch, an optical signal is used to switch the path another optical signal takes through the switch.

Performance

Various parameters are defined and specified to quantify the performance of optical switches. The steady state performance of an optical switch (or optical switching matrix) is measured by its ability to effectively transmit optical power from an input port to any one of N output ports over the “on” state transmission path, and its ability to effectively isolate input power sources from all non-active ports over the “off” state transmission paths. Other key optical performance parameters include transmission efficiency over a range of wavelengths, the ability to minimize input optical power reflected back into the input fiber, transmission balance, and bidirectional transmission. The optical switch (or switching matrix) transient behavior is another important characteristic that is specified by its speed of response to control stimulation via the time interval it takes to either transmit or block the optical signal on any given output port.
Two rates can be associated with switches: the switching rate and the signal transmission rate. The switching rate is the rate at which a switch changes states. The signal transmission rate is the modulation rate of information passing through a switch. The signal transmission rate is usually much greater than the switching rate. (If the switching rate approaches or exceeds the transmission rate, then the switch can be called an optical modulator.)
A switch’s ability to sustain its steady state and transient performance specifications under stressful environmental conditions and over time is also an important characteristic.

Applications

Optical switching technology is driven by the need to provide flexibility in optical network connectivity. Prime applications are optical protection, test systems, remotely reconfigurable add-drop multiplexers, and sensing. Possible future applications include remote optical provisioning and restoration.
Current switching applications include passive protection switching for service restoration following a disruption, such as a fiber cut. One common application for switches is in Remote Fiber Test Systems (RFTSs) that can monitor and locate a fault on a fiber transmission line. An emerging application of optical switches is optical cross-connection. Optical cross-connects utilize optical switching fabrics to establish an interconnection between multiple optical inputs and outputs.



__________________________________________________________________________________


                         Optical Circuits: Single Photon Flips Transistor Switch


Photons emerge as competitors to electrons in new computer circuits

Transistors, the tiny switches that flip on and off inside computer chips, have long been the domain of electricity. But scientists are beginning to develop chip components that run on light. Last week, in a remarkable achievement, a team led by researchers at the Massachusetts Institute of Technology (MIT) in Cambridge reported building a transistor that is switched by a single photon.
Conventionally, photons are used only to deliver information, racing along fiber-optic cables with unparalleled speed. The first commercial silicon chip to include optical elements, announced last December, did little to challenge the status quo. The on-board beams of light in the device, developed at IBM’s research center in Yorktown Heights, New York, merely shuttle data between computer chips.
Now, Wenlan Chen of MIT and her colleagues have taught light some new tricks, using a cloud of chilled caesium atoms suspended between two mirrors. Their transistor is set to ‘on’ by default, allowing a beam of light to sail through the transparent caesium cloud unmolested. But sending in a single ‘gate’ photon turns the switch off, thanks to an effect called electromagnetically induced transparency. The injected photon excites the caesium atoms, rendering them reflective to light trying to cross the cloud (see ‘Turn off the light’). One photon can thus block the passage of about 400 other photons, says Chen, who presented the result on 7 June at a meeting of the American Physical Society’s Division of Atomic, Molecular and Optical Physics in Quebec City, Canada.

The ability to turn a strong signal on and off using a weak one fulfills a key requirement of an optical transistor. “Nothing even came close before,” says physicist Ataç mamolu of the Swiss Federal Institute of Technology Zurich, who called the experiment “a true breakthrough”. In theory, the hundreds of photons, controlled by the triggering photon, could fan out and switch off hundreds of other transistors in an optical circuit.
With its exotic clouds of atoms and bulky equipment, the proof-of-principle transistor is unlikely to become a component in everyday computers. But it could be a useful tool for studying how photons interact at the quantum level — potentially leading to a quantum transistor that flips, not a one or a zero as in classical computing, but a fuzzy bit of quantum information.
A more practical optical transistor debuted in April 2012 at Purdue University in West Lafayette, Indiana, where electrical engineer Minghao Qi has made one that is compatible with the semiconductor industry’s existing manufacturing techniques. “The advantage of our device is that we have it on a silicon chip,” says Qi.
In this case, the beam of light to be switched on and off enters and exits along a channel, etched in the silicon, that sits next to a parallel channel. In between the two rails is an etched ring. When a weaker light beam courses through the second optical line, the ring heats up and swells, interfering with the main beam and switching off the transistor. This switch can flip on and off up to 10 billion times per second.

And the output beam can fan out and drive two other transistors, meeting one of the established requirements for an optical transistor set out in 2010 by David Miller, a physicist at Stanford University in California. Other cri­teria include matching the frequency of the exiting signal to the input frequency and keeping the output clean, with no degradation that could cause errors. “Making an optical transistor that really satisfies the necessary criteria is very hard,”

Still, Qi does not expect to challenge the electronic transistor with his optical analogue, which consumes a lot more power and runs much more slowly. “We want to complement the Intel transistor,” he says. “We don’t want to replace it.” He hopes to find a foothold in niche markets, such as equipment for scrambling cable channels and military technologies that could benefit from light’s imperviousness to an electromagnetic attack.
Rout

ers that guide information through the Internet could also be amenable to optical transistors and switches. At present, these stopping points in the network convert optical signals travelling through fiber-optic cables into electrical signals; these are then processed, converted back to light and sent on their way. A router in which one beam of light pushes another in the appropriate direction — with no conversions involved — could in principle be faster and consume less energy.

A popular candidate for such switches are quantum dots, small semiconductor crystals that behave like atoms. In one particularly sensitive quantum-dot switch, a beam of light is first guided along a material dotted with holes, called a photonic crystal. The light can pass through a quantum dot placed in its path without changing course. But if a pulse of light is sent in just ahead of that beam, it can induce an inter­action between the dot and the crystal that scatters the beam and sends it on a different path.


But the switch still faces a practical obstacle common to all of these emerging optical technologies. The lasers that supply the devices with light consume considerable energy, offsetting any savings. “Right now,”

      Gambar terkait

_________________________________________________________________________________

                        Applied Robotics: How Robots Are Changing Our World


Photonics and artificial intelligence are enabling the creation of robots with new ways of interacting in business, medicine, imaging and many other applications.

The robot age is upon us. The thought of robots might bring to mind helpful androids like C-3PO in “Star Wars” or Rosie from “The Jetsons” – or it might stir up concern for humankind as advanced robots become more and more indispensable and take over dangerous or boring tasks. Either way, most people don’t realize how ubiquitous robots already are because in their most common forms today, they are less android and more like industrial equipment or tools. Incorporating photonics technology like cameras, sensors, lasers, displays and facial recognition technology, robotics are found everywhere from industrial processing to devices such as pool cleaners and Google’s self-driving car.

Advanced robotics can be found in important commercial applications, such as food processing, packaging and industrial control, often in the form of a programmable, automated “arm.” Consumer robots like the Roomba vacuum cleaner and the Mirra pool cleaner from iRobot Corp. in Bedford, Mass., are not even vaguely humanoid but help with everyday household tasks. The Ava 500 from iRobot is a business-class telepresence robot that can maneuver automatically through hallways and manufacturing floors to enable remote collaboration (Figure 1).


The Ava 500 telepresence robot from iRobot has completed field trials to facilitate collaboration across a distance via intelligent and safe self-navigation to preselected destinations.
Figure 1. The Ava 500 telepresence robot from iRobot has completed field trials to facilitate collaboration across a distance via intelligent and safe self-navigation to preselected destinations. The user can manually rotate, move up or down, or tilt the robot to image the environment in high-definition video. Photo courtesy of iRobot.

Military robots provide critical search and data-collection functions above ground and underwater, reporting hazards to keep forces safe. Robotic remote-controlled boats and vehicles might serve as educational and entertainment tools in the hobbyist space, or as payload delivery devices, as in the case of unmanned drones. Scientists have developed robots such as the Mars rovers for remote exploration and the da Vinci Surgical System for performing delicate operations (Figure 2).


The da Vinci Surgical System was the first robotic surgery system approved by the FDA for general laparoscopic surgery.
Figure 2. The da Vinci Surgical System was the first robotic surgery system approved by the FDA for general laparoscopic surgery. The physician uses a 3-D high-definition vision system to operate with a 10× magnified view, enabling enhanced precision, dexterity and control. Photo courtesy of da Vinci Surgery.

According to the International Federation of Robotics (IFR), 2013 was the biggest year ever for global sales of industrial robots in the automotive, chemical, materials and food processing industries. The automotive sector uses one-third of all industrial robots, including those to help manufacture cars, as well as robotic technology to help park and control vehicles and avoid collision. The IFR estimates that from 2012 to 2013, the global demand for personal and domestic service robots grew 28 percent to $1.7 billion. Handicap assistance robots are a burgeoning class, as are robotic systems for gaming.

The classic form of robots has made amazing progress, too. With the announcement in January that a hotel in Japan would be staffed by eerily humanlike but entirely robotic personnel, the idea of fleshy humanoid robots from the “Futureworld” or “Westworld” movies just got a step closer (Figure 3). The Henn-na Hotel in Nagasaki Prefecture, which translates as “Strange Hotel,” will have receptionist robots with strong human likenesses that will greet visitors and engage in intelligent conversations. The hotel, scheduled to open in July with 72 rooms, is part of a theme park called Huis Ten Bosch, modeled after a royal palace in the Netherlands. Robots will also provide room service, porter service and housekeeping. Huis Ten Bosch President Hideo Sawada hopes the hotel will be made so efficient through the use of robots and automated features that its success will enable hundreds of hotels like it around the world.


 Robots with human-like flesh, robotic vision and conversation skills via artificial intelligence will greet guests in the reception area of the Henn-na Hotel, planned to open this year in Nagasaki, Japan.
Figure 3. Robots with human-like flesh, robotic vision and conversation skills via artificial intelligence will greet guests in the reception area of the Henn-na Hotel, planned to open this year in Nagasaki, Japan. The kiosk-assisted check-in will use facial recognition technology instead of keys to allow guests access to their rooms. Photo courtesy of seejapan.co.uk.


Less humanoid but more tactile and logic oriented is the Baxter series of robots conceived at Rethink Robotics Inc. in Boston. Baxter is an interactive humanoid robot platform that incorporates 360° sonar sensors, a 1024 × 600 SVGA LCD display “head” and advanced, customized software. Baxter operates beside existing skilled workers to optimize manufacturing, packaging and research processes (Figures 4 and 5). Baxter’s cameras support computer vision applications with 30-fps image capture rate and an effective resolution of 640 × 400 pixels.

The Baxter series of robots offers imaging technology, advanced sensor design and avoidance response that enable the robots to work safely beside humans in industrial environments.

Figure 4. The Baxter series of robots offers imaging technology, advanced sensor design and avoidance response that enable the robots to work safely beside humans in industrial environments. Photo courtesy of Rethink Robotics.

At the official launch of the Australian Centre for Robotic Vision (ACRV) in Brisbane, Australia, in March, researchers from the Queensland University of Technology introduced a Baxter model programmed to use computer vision to play an unbeatable game of Connect Four. They also are programming the robot to pick ripe bell peppers according to color.

“Robotic vision is the key enabling technology that will allow robotics to transform labor-intensive industries, disrupt stagnant markets and see robots become a ubiquitous feature of the modern world,” said Sue Keay, chief operating officer at ACRV.
Figure 5. Queensland University of Technology unveiled a version of a Baxter robot that is programmed to identify and pick ripe bell peppers, as well as to play an unbeatable game of Connect Four. Photo courtesy of ACRV.

Figure 5. Queensland University of Technology unveiled a version of a Baxter robot that is programmed to identify and pick ripe bell peppers, as well as to play an unbeatable game of Connect Four. Photo courtesy of ACRV.

An IR sensor with a range of 1.5 to 15 in. (4 to 40 cm) is part of Baxter’s accidental-contact avoidance system. The ability to recognize obstacles suddenly in their way and to avoid them is an important feature of robots’ ability to work safely beside humans.

“We’ve tried to ‘rush’ Baxter before and can vouch for his quick reflexes,” said roboticist Peter Corke, professor of science and engineering at Queensland University of Technology and director of ACRV. The center plans to tackle the problem of robot vision by designing new types of low-cost sensors to improve image quality, as well as new algorithms that will accommodate imaging even in poor conditions, such as in the dark or in extreme temperatures, cued by context and sensing of the environment.

In April, Rethink Robotics announced the completion of a Series D round of funding for $40 million, bringing total investment in the company to $113.5 million since its founding in 2008. Said Scott Eckert, president and CEO, “A shortage of manufacturing labor around the world, coupled with manufacturers’ need to respond rapidly to market and product changes, is creating the need for a new kind of robot.”

Flying robots

Another kind of robot is about to become a much bigger part of our world: the unmanned aerial vehicle (UAV) or drone. Drones, like many other robotic technologies, go hand in hand with photonics. Where drones go, cameras, sensors and imaging technology can go. Amazon.com plans to use small drones for package delivery in 30 minutes or less. This plan took another step ahead in March, when the FAA issued permission for Amazon to conduct testing during the daytime within sight of a pilot with a private pilot’s certificate. Also in March, Facebook announced plans to bring Internet connectivity to remote areas using a network of drones and satellites connected via IR lasers in a free-space optics network. Facebook acquired solar-powered drone company Ascenta (UK) Ltd., based in Somerset, UK, and talent from NASA to tackle the project. In April, Google acquired high-altitude drone company Titan Aerospace Corp. of Moriarty, N.M., whose drones have the capability to provide high-quality images in real time that could map the Earth from 65,000 feet high for up to three years while contributing to disaster relief and addressing deforestation. Clearly, these titans of industry recognize the potential of drone technology and are making big investments to make sure that flying robots become a commonplace sight in our world very soon.

That is, if they can get through the bottleneck that is the U.S. Federal Aviation Administration. For the FAA, safely enabling high-volume deployment of private and commercial drones to operate within the regulated environment of the military, commercial and private airspaces is a complex, time-consuming chore. In March, however, the FAA proposed a long-awaited regulatory framework that provides a simple process for drone users applying for exemptions to allow small drones in certain airspaces.

“With this exemption process in play, the market is now open,” says Bryan da Frota, CEO of Prioria Robotics Inc. of Gainesville, Fla., maker of the Maveric UAS, a portable and rapidly deployable unmanned vehicle. In November 2013, Prioria announced several contracts totaling $4.5 million to provide the U.S. Army Rapid Equipping Force with the Maveric Unmanned Aircraft System, a single-person operable and portable aircraft with numerous payload options. Maveric has foldable wings and can be launched from a portable tube. These features enable the system to be carried and hand launched for imaging within 10 miles of a ground station. In some applications, the Maveric can provide resolution of 1 cm per pixel at 300 feet, making it suitable for defense imaging and for surveying in both the mining and oil and gas industries (Figure 6).


The Maveric from Prioria Robotics is a backpack-able, hand-launchable UAV platform measuring just over 2 ft. in wingspan that can carry a wide variety of imaging and sensing payloads for civilian (<b>left</b>) and military (<b>right</b>) applications.
Figure 6. The Maveric from Prioria Robotics is a backpack-able, hand-launchable UAV platform measuring just over 2 ft. in wingspan that can carry a wide variety of imaging and sensing payloads for civilian (left) and military (right) applications. Photo courtesy of Prioria Robotics.


Prioria is working on applying image processing to usher in the next generation of robotic capability.

“How does a UAV fly autonomously in an area without GPS?” said da Frota. “How does it avoid obstacles and edit the payload data before transmission so it doesn’t overwhelm the data links with information that isn’t relevant?”

Maveric is equipped to carry a variety of payloads, from simple electro-optical color cameras and IR thermal imaging cameras, to near-IR and short-wave IR systems that allow for agricultural imaging of soil moisture, plant health and imaging through smoke. It also can carry gas sensors and particulate detectors, and eventually the company hopes it will carry lasers. Prioria is installing a hyperspectral imaging payload with an agricultural partner, and in May, the company announced an industrial-grade mapping pod that is the first of its size, quality and class. For these and many other applications, drones and other robotic technology seem poised to become part of our everyday world.

    Hasil gambar untuk photons in electronic equipment





___________________________________________________________________________________


      Gambar terkait


                           Photons on a chip set new paths for secure communications


                  Gambar terkait

                   Integrated photonic qubit quantum computing on a superconducting chip

RMIT research has helped crack the code to ultra-secure telecommunications of the future in an international research project that could also expedite the advent of quantum computing.


A team co-led by RMIT MicroNano Research Facility Director Professor David Moss has added a new twist to create that fit on a tiny computer chip.
The breakthrough, published in Nature Communications, heralds the next-generation of integrated quantum optical technology, being compatible with current technology and secure communications.
The team pioneered a new approach based on a micro-ring resonator - a tiny optical cavity - in which energy conservation constraints can be exploited to suppress classical effects while amplifying quantum processes.
They used laser beams at different wavelengths and then had to overcome the risk of the two pump beams being able to destroy the photons' fragile quantum state.
"One of the properties of light exploited within quantum optics is 'photon polarization', which is essentially the direction in which the electric field associated with the photon oscillates,'' Moss said.
"Processes used to generate single photons or photon pairs on a chip allow the generation of photons with the same polarization as the laser beam, forcing us to find a way to directly mix, or cross-polarize, the photons via a nonlinear optical process on a chip for the first time.''
Moss worked with Professor Roberto Morandotti at the INRS-EMT in Canada and researchers from the University of Sussex and Herriot Watt University, City University of Hong Kong, and the Xi'an Institute in Chin, on the research.
"While a similar suppression of classical effects has been observed in gas vapours and complex micro-structured fibres, this is the first time it has been reported on a chip, opening a route for building scalable integrated devices that exploit the mixing of polarization on a single level,'' he said.
"It also has the advantage that the fabrication process of the chip is compatible with that currently used for electronic chips which not only allows the exploitation of the huge global infrastructure of CMOS foundries, but will ultimately offer the potential to integrate electronic devices on the same chip.
"Both of these are fundamental requirements for the ultimate widespread adoption of optical quantum technologies.''

                                        Gambar terkait
   






     _____________________________________________________________________________

                                      SPACE ENGINEERING AND TECHNOLOGY


  
                         
                           
                                             Optical Communication Facility




               MECHANTRONICS AND OPITICS ___ ROBOTICS AND LIFE SUPPORT 


The ability to move and see is every bit as useful for space hardware as it is for their human makers.

Herschel - View of optical mirrors
 
Mechatronics is the fusion of mechanical, electrical, optical and opto-electronical, material and bio-technology systems. It contributes to the development of advanced robotics, of instrumentation for physical or life sciences, of optical instruments for remote sensing, of devices which transmit and  detect light for communication or processing, as well as for the development of life support systems.

RAT rover by night
 
Planetary rovers and “robotic-arm” systems are the highest profile applications of robotics in space. Mobile robots designed to explore celestial bodies on or under their surfaces are on the front line of space exploration while robotic arms can perform challenging inspection and servicing duties previously undertaken by human spacewalkers. 

Just as important in terms of scientific return are the kind of sophisticated experiment payloads hosted on ESA's Columbus module – such as furnaces for crystal growth and fluid science facilities for the physical sciences. On the life science side, the equipment developed includes cultivation and handling devices, contamination control (including microbial contamination) and precision cleaning as well as micro-sensors for analysis and diagnosis.

Life Support and Physical Sciences Laboratory
 
The design and verification of optical systems is another core activity, ranging from full-sized telescopes to fibre optics and photonics devices, from  laser communications  to lidar atmosphere-sampling sensors and space interferometers. This work extends into advanced opto-electronic systems such as superconducting magnetic field detection devices, advanced detectors operating across a broad spectral range from X-rays and gamma rays into the infrared to the application of  advanced quantum states  for secure communications.
Not to be forgotten, the preparation of advanced  regenerative life support system needed for long term journey and stay on the Moon or Mars, including all air / water / food reclaim aspects and the related safety issues.






__________________________________________________________________________________


Gen. Mac TECH zone MARIA PREFER in electronic components traveling a few light years will use photons in future electronic equipment to replace the speed of the switching process and signal conditioner in the form of old electronic components such as silicon and germanium microcontrolers and nanocontrolers 

        
         Hasil gambar untuk robotics and photonics  Gambar terkait
__________________________________________________________________________________




Tidak ada komentar:

Posting Komentar