Jumat, 20 Maret 2020

JESI ISE Loving 2 in 1 for the placement of modern humans and humanoid robots outside the earth by removing all the earth's gravity systems in the EINSTEIN area to the point of harmonization G4 along with energy placement techniques that are more durable, large areas, and always synchronizing with places in the external system space such as: solar energy, cosmic energy, star energy, the use of space and time on black hole energy, space mining special research organs on planetary systems, mastery of extraterrestrial energy; welcome to the future structure of space energy research resources and transverse angel wings AMNIMARJESLO SAN 1244 wave guide-- Gen. Mac Tech making 2 directions into 1 space fighter aircraft energy technique __ open track to sky heaven POSTING




                 Hasil gambar untuk light speed  _____  Hasil gambar untuk light speed


when the population of the earth increased in early humans in the 20th and 21st century it only expects energy sources based on natural systems from the effects of I / O sources of gravity such as: alternating energy sources that occur due to gravity's gravitational pull system namely the north and south poles, wind energy systems are also formed from Earth's gravity, all mining materials are also influenced by the natural formation of the Earth's gravity theory which can be explained by EINSTEIN's postulates on G-Force 1, 2, 3, 4. although at some point EINSTEIN is honored to place the speed of light at E = MC ^ 2. because light is formed not from Earth's gravity but from a state of emptiness and nothingness in outer space so that empty space activates its radiation system to fill nothingness for a long time. when SUPERMAN observes outer space he is only a small beam between many radiation systems and modifiers, SUPERMAN requires SUPERGIRL to measure the distance of light and the speed of light transverse radiation while we are on earth under the influence of Earth's gravity so the measured is still below the standard speed. speed is directly proportional to the amount of energy. 


                                                                                  
                                                            Sincerely , 
                                                AMNIMARJESLO  SAN

            
                                Hasil gambar untuk light speed

    


          (  Gen. Mac Tech making 2 directions into 1 space fighter aircraft energy technique )

                                           Hasil gambar untuk spaceship with starlight energy

                     _____e- Solar Energy_____



Solar energy is the radiant light and heat from the sun that has been harnessed by humans since ancient times using a range of ever-evolving technologies. Solar radiation along with secondary solar resources account for most of the available renewable energy on earth.

because solar energy is used for needs on earth, it is usually used to be an AC supply component (alternating current), which uses an electronic INVERTER circuit and a DC supply component uses an accumulator. whereas outside the earth there will be an intersection of space and time on planets outside the earth because space and time are different so the energy graph and formulation system must be adjusted to the e-STAR C located there . 

Solar energy is radiant light and heat from the Sun that is harnessed using a range of ever-evolving technologies such as solar heating, photovoltaics, solar thermal energy, solar architecture, molten salt power plants and artificial photosynthesis.
It is an important source of renewable energy and its technologies are broadly characterized as either passive solar or active solar depending on how they capture and distribute solar energy or convert it into solar power. Active solar techniques include the use of photovoltaic systems, concentrated solar power and solar water heating to harness the energy. Passive solar techniques include orienting a building to the Sun, selecting materials with favorable thermal mass or light-dispersing properties, and designing spaces that naturally circulate air.
The large magnitude of solar energy available makes it a highly appealing source of electricity. The United Nations Development Programme in its 2000 World Energy Assessment found that the annual potential of solar energy was 1,575–49,837 exajoules (EJ). This is several times larger than the total world energy consumption, which was 559.8 EJ in 2012.
In 2011, the International Energy Agency said that "the development of affordable, inexhaustible and clean solar energy technologies will have huge longer-term benefits. It will increase countries’ energy security through reliance on an indigenous, inexhaustible and mostly import-independent resource, enhance sustainability, reduce pollution, lower the costs of mitigating global warming, and keep fossil fuel prices lower than otherwise. These advantages are global. Hence the additional costs of the incentives for early deployment should be considered learning investments; they must be wisely spent and need to be widely shared .

Solar irradiation data is needed at all levels of solar power development, from initial government planning through to large-scale project development or the calculations needed to size smaller systems. In the past such data was provided at a relatively course level from NASA and other global providers, but more recently specialist models have been developed to more precisely calculate global horizontal irradiation (GHI) and direct normal irradiation (DNI) using primarily cloud cover data from satellites. A number of firms now offer such data as a commercial service. Based on this, it is possible to calculate average annual power output from a theoretical photovoltaic power plant (PVOUT), taking into account temperature, tilt, and the efficiency of the equipment being used (solar panels and balance of system components).

 Solar energy technologies refer primarily to the use of solar radiation for practical ends. All other renewable energies other than geothermal derive their energy from energy received from the sun.
Solar technologies are broadly characterized as either passive solar or active solar depending on the way they capture, convert and distribute sunlight. Active solar techniques include the use of photovoltaic modules (also called photovoltaic panels) and solar thermal collectors (with electrical or mechanical equipment) to convert sunlight into useful outputs. Passive solar techniques include orienting a building to the Sun, selecting materials with favorable thermal mass or light dispersing properties, and designing spaces that naturally circulate air.
Active solar technologies increase the supply of energy and are considered supply side technologies, while passive solar technologies reduce the need for alternate resources and are generally considered demand side technologies .

Solar energy is the most abundant, renewable energy source in the world. Solar energy systems refer to technologies that convert the sun's heat or light to another form of energy for use.
There are two categories of technologies that harness solar energy, Solar Photovoltaics and Solar Thermal. Solar Photovoltaic (or PV) is a technology that converts sunlight into direct current electricity by using semiconductors. In contrast, Solar Thermal is a technology that utilizes the heat energy from the sun for heating or electricity production .

Solar is a renewable resource and does not emit any greenhouse gases in the energy generation process. However, the cost of solar in relation to other fuel sources is a barrier to adoption. Additionally, sunlight varies depending on geographic location, season, and time of day, which all creates limitations on its use.


                                       
                                              


                                                     
                                              
       
    

                          Solar Power from Satellites

 
This drawing shows how power collected by solar-power satellites might be beamed to a receiving antenna on Earth. The antenna would convert the energy back into electrical current, which would be fed into the power grid. Courtesy: Space Studies Institute.
The sun powers the biosphere, which is to say that the energy used by almost all plants and animals comes from the sun. So why not use solar energy to power industry, transportation, and the home as well? Well, a principal difficulty with solar power is that the sun doesn't always shine on a particular location: half the time the earth blocks the sun, and for much of the remaining time clouds and fog do. But what if the solar energy were collected by a set of satellites above the earth’s atmosphere? Then we might obtain solar power for 24 hours every day of the year. This is the idea behind solar-power satellites.

A satellite with solar panels to convert light energy into electricity can be put into orbit. Indeed, most satellites in orbit today are powered by solar panels. But how can we get the energy from the satellite back to earth? Clearly it would be impossible to use the electric lines we use for long-distance power transmission on earth. This is where microwaves come in.

The idea is that a satellite be equipped with a microwave generator, so that the electrical energy from the solar panels can be converted into a microwave beam. Then the microwave beam can be directed to antennas on the surface of the earth, which would convert the microwaves back to electrical energy. The energy could then either be used at the site of the antenna or injected into the electric-power network.

It was during the late 1960s that the engineer Peter Glaser first had the notion of solar power satellites. The principle of transmitting power by microwaves had already been demonstrated, though not put into practice. (Microwaves in practical devices, such as radar systems and long-distance telephone relays, were used to convey information.) To convey information, the intensity of the received signal need only be less than one nanowatt (one billionth of a watt). Glaser’s idea was to put the solar-power satellites in geosynchronous orbits, so that each would hover over a single location on the earth. This meant, however, that the satellites had to be very high (36,000 kilometers or about 22,000 feet), and this in turn meant that the antenna on the satellite and the receiving antenna on the ground had to be extremely large (a kilometer or more in diameter). The idea did not seem practical, and after some initial funding by the U.S Department of Energy and NASA there was little interest in pursuing the technology.

Today, however, the situation is changed because of the very large number of communications satellites in low orbits. It might be possible to make these satellites dual purpose—solar-energy collectors as well as communications devices. Because of the much lower orbits, the antennas on the satellites and on the ground need not be nearly so large. A drawback however, is that satellites in low-earth orbit circle the earth rapidly (about every 90 minutes) and therefore do not provide a connection for a very long time. There are also other concerns. One is that the transmission down to the ground might be interrupted by clouds and weather. Another is the safety of the people and animals near the receiving antennas who might be exposed to the microwave radiation. Today, the viability of solar-power satellites as a long-term solution to our energy needs is being investigated by government agencies and individual companies in many countries.


                                                     🔀   Power electronics ⬲


Power electronics is the application of solid-state electronics to the control and conversion of electric power.

The first high power electronic devices were mercury-arc valves. In modern systems, the conversion is performed with semiconductor switching devices such as diodes, thyristors, and power transistors such as the power MOSFET and IGBT. In contrast to electronic systems concerned with transmission and processing of signals and data, in power electronics substantial amounts of electrical energy are processed. An AC/DC converter (rectifier) is the most typical power electronics device found in many consumer electronic devices, e.g. television sets, personal computers, battery chargers, etc. The power range is typically from tens of watts to several hundred watts. In industry a common application is the variable speed drive (VSD) that is used to control an induction motor. The power range of VSDs start from a few hundred watts and end at tens of megawatts.
The power conversion systems can be classified according to the type of the input and output power
                                                  
          A PCs power supply is an example of a piece of power electronics, whether inside or outside of the cabinet 


DC to AC converters produce an AC output waveform from a DC source. Applications include adjustable speed drives (ASD), uninterruptible power supplies (UPS), Flexible AC transmission systems (FACTS), voltage compensators, and photovoltaic inverters. Topologies for these converters can be separated into two distinct categories: voltage source inverters and current source inverters. Voltage source inverters (VSIs) are named so because the independently controlled output is a voltage waveform. Similarly, current source inverters (CSIs) are distinct in that the controlled AC output is a current waveform.
DC to AC power conversion is the result of power switching devices, which are commonly fully controllable semiconductor power switches. The output waveforms are therefore made up of discrete values, producing fast transitions rather than smooth ones. For some applications, even a rough approximation of the sinusoidal waveform of AC power is adequate. Where a near sinusoidal waveform is required, the switching devices are operated much faster than the desired output frequency, and the time they spend in either state is controlled so the averaged output is nearly sinusoidal. Common modulation techniques include the carrier-based technique, or Pulse-width modulation, space-vector technique, and the selective-harmonic technique.
Voltage source inverters have practical uses in both single-phase and three-phase applications. Single-phase VSIs utilize half-bridge and full-bridge configurations, and are widely used for power supplies, single-phase UPSs, and elaborate high-power topologies when used in multicell configurations. Three-phase VSIs are used in applications that require sinusoidal voltage waveforms, such as ASDs, UPSs, and some types of FACTS devices such as the STATCOM. They are also used in applications where arbitrary voltages are required as in the case of active power filters and voltage compensators.[15]
Current source inverters are used to produce an AC output current from a DC current supply. This type of inverter is practical for three-phase applications in which high-quality voltage waveforms are required.
A relatively new class of inverters, called multilevel inverters, has gained widespread interest. Normal operation of CSIs and VSIs can be classified as two-level inverters, due to the fact that power switches connect to either the positive or to the negative DC bus. If more than two voltage levels were available to the inverter output terminals, the AC output could better approximate a sine wave. It is for this reason that multilevel inverters, although more complex and costly, offer higher performance.
Each inverter type differs in the DC links used, and in whether or not they require freewheeling diodes. Either can be made to operate in square-wave or pulse-width modulation (PWM) mode, depending on its intended usage. Square-wave mode offers simplicity, while PWM can be implemented several different ways and produces higher quality waveforms.
Voltage Source Inverters (VSI) feed the output inverter section from an approximately constant-voltage source.
The desired quality of the current output waveform determines which modulation technique needs to be selected for a given application. The output of a VSI is composed of discrete values. In order to obtain a smooth current waveform, the loads need to be inductive at the select harmonic frequencies. Without some sort of inductive filtering between the source and load, a capacitive load will cause the load to receive a choppy current waveform, with large and frequent current spikes.
There are three main types of VSIs:
  1. Single-phase half-bridge inverter
  2. Single-phase full-bridge inverter
  3. Three-phase voltage source inverter

Single-phase half-bridge inverter

Figure 8: The AC input for an ASD.
FIGURE 9: Single-Phase Half-Bridge Voltage Source Inverter
The single-phase voltage source half-bridge inverters, are meant for lower voltage applications and are commonly used in power supplies. Figure 9 shows the circuit schematic of this inverter.
Low-order current harmonics get injected back to the source voltage by the operation of the inverter. This means that two large capacitors are needed for filtering purposes in this design. As Figure 9 illustrates, only one switch can be on at time in each leg of the inverter. If both switches in a leg were on at the same time, the DC source will be shorted out.
Inverters can use several modulation techniques to control their switching schemes. The carrier-based PWM technique compares the AC output waveform, vc, to a carrier voltage signal, vΔ. When vc is greater than vΔ, S+ is on, and when vc is less than vΔ, S- is on. When the AC output is at frequency fc with its amplitude at vc, and the triangular carrier signal is at frequency fΔ with its amplitude at vΔ, the PWM becomes a special sinusoidal case of the carrier based PWM. This case is dubbed sinusoidal pulse-width modulation (SPWM).For this, the modulation index, or amplitude-modulation ratio, is defined as ma = vc/v .
The normalized carrier frequency, or frequency-modulation ratio, is calculated using the equation mf = f/fc .
If the over-modulation region, ma, exceeds one, a higher fundamental AC output voltage will be observed, but at the cost of saturation. For SPWM, the harmonics of the output waveform are at well-defined frequencies and amplitudes. This simplifies the design of the filtering components needed for the low-order current harmonic injection from the operation of the inverter. The maximum output amplitude in this mode of operation is half of the source voltage. If the maximum output amplitude, ma, exceeds 3.24, the output waveform of the inverter becomes a square wave.
As was true for Pulse Width Modulation (PWM), both switches in a leg for square wave modulation cannot be turned on at the same time, as this would cause a short across the voltage source. The switching scheme requires that both S+ and S- be on for a half cycle of the AC output period. The fundamental AC output amplitude is equal to vo1 = vaN = 2vi.
Its harmonics have an amplitude of voh = vo1/h.
Therefore, the AC output voltage is not controlled by the inverter, but rather by the magnitude of the DC input voltage of the inverter.
Using selective harmonic elimination (SHE) as a modulation technique allows the switching of the inverter to selectively eliminate intrinsic harmonics. The fundamental component of the AC output voltage can also be adjusted within a desirable range. Since the AC output voltage obtained from this modulation technique has odd half and odd quarter wave symmetry, even harmonics do not exist. Any undesirable odd (N-1) intrinsic harmonics from the output waveform can be eliminated.

Single-phase full-bridge inverter

FIGURE 3: Single-Phase Voltage Source Full-Bridge Inverter
FIGURE 4: Carrier and Modulating Signals for the Bipolar Pulsewidth Modulation Technique
The full-bridge inverter is similar to the half bridge-inverter, but it has an additional leg to connect the neutral point to the load. Figure 3 shows the circuit schematic of the single-phase voltage source full-bridge inverter.
To avoid shorting out the voltage source, S1+ and S1- cannot be on at the same time, and S2+ and S2- also cannot be on at the same time. Any modulating technique used for the full-bridge configuration should have either the top or the bottom switch of each leg on at any given time. Due to the extra leg, the maximum amplitude of the output waveform is Vi, and is twice as large as the maximum achievable output amplitude for the half-bridge configuration.
States 1 and 2 from Table 2 are used to generate the AC output voltage with bipolar SPWM. The AC output voltage can take on only two values, either Vi or –Vi. To generate these same states using a half-bridge configuration, a carrier based technique can be used. S+ being on for the half-bridge corresponds to S1+ and S2- being on for the full-bridge. Similarly, S- being on for the half-bridge corresponds to S1- and S2+ being on for the full bridge. The output voltage for this modulation technique is more or less sinusoidal, with a fundamental component that has an amplitude in the linear region of less than or equal to one vo1 =vab1= vi • ma.
Unlike the bipolar PWM technique, the unipolar approach uses states 1, 2, 3 and 4 from Table 2 to generate its AC output voltage. Therefore, the AC output voltage can take on the values Vi, 0 or –V [1]i. To generate these states, two sinusoidal modulating signals, Vc and –Vc, are needed, as seen in Figure 4.
Vc is used to generate VaN, while –Vc is used to generate VbN. The following relationship is called unipolar carrier-based SPWM vo1 =2 • vaN1= vi • ma.
The phase voltages VaN and VbN are identical, but 180 degrees out of phase with each other. The output voltage is equal to the difference of the two phase voltages, and do not contain any even harmonics. Therefore, if mf is taken, even the AC output voltage harmonics will appear at normalized odd frequencies, fh. These frequencies are centered on double the value of the normalized carrier frequency. This particular feature allows for smaller filtering components when trying to obtain a higher quality output waveform.
As was the case for the half-bridge SHE, the AC output voltage contains no even harmonics due to its odd half and odd quarter wave symmetry.

Three-phase voltage source inverter

FIGURE 5: Three-Phase Voltage Source Inverter Circuit Schematic
FIGURE 6: Three-Phase Square-Wave Operation a) Switch State S1 b) Switch State S3 c) S1 Output d) S3 Output
Single-phase VSIs are used primarily for low power range applications, while three-phase VSIs cover both medium and high power range applications. Figure 5 shows the circuit schematic for a three-phase VSI.
Switches in any of the three legs of the inverter cannot be switched off simultaneously due to this resulting in the voltages being dependent on the respective line current's polarity. States 7 and 8 produce zero AC line voltages, which result in AC line currents freewheeling through either the upper or the lower components. However, the line voltages for states 1 through 6 produce an AC line voltage consisting of the discrete values of Vi, 0 or –Vi.
For three-phase SPWM, three modulating signals that are 120 degrees out of phase with one another are used in order to produce out of phase load voltages. In order to preserve the PWM features with a single carrier signal, the normalized carrier frequency, mf, needs to be a multiple of three. This keeps the magnitude of the phase voltages identical, but out of phase with each other by 120 degrees.The maximum achievable phase voltage amplitude in the linear region, ma less than or equal to one, is vphase = vi / 2. The maximum achievable line voltage amplitude is Vab1 = vab • 3 / 2
The only way to control the load voltage is by changing the input DC voltage.

Current source inverters

FIGURE 7: Three-Phase Current Source Inverter
Figure 8: Synchronized-Pulse-Width-Modulation Waveforms for a Three-Phase Current Source Inverter a) Carrier and Modulating Signals b) S1 State c) S3 State d) Output Current
Figure 9: Space-Vector Representation in Current Source Inverters
Current source inverters convert DC current into an AC current waveform. In applications requiring sinusoidal AC waveforms, magnitude, frequency, and phase should all be controlled. CSIs have high changes in current over time, so capacitors are commonly employed on the AC side, while inductors are commonly employed on the DC side. Due to the absence of freewheeling diodes, the power circuit is reduced in size and weight, and tends to be more reliable than VSIs. Although single-phase topologies are possible, three-phase CSIs are more practical.
In its most generalized form, a three-phase CSI employs the same conduction sequence as a six-pulse rectifier. At any time, only one common-cathode switch and one common-anode switch are on.
As a result, line currents take discrete values of –ii, 0 and ii. States are chosen such that a desired waveform is output and only valid states are used. This selection is based on modulating techniques, which include carrier-based PWM, selective harmonic elimination, and space-vector techniques.
Carrier-based techniques used for VSIs can also be implemented for CSIs, resulting in CSI line currents that behave in the same way as VSI line voltages. The digital circuit utilized for modulating signals contains a switching pulse generator, a shorting pulse generator, a shorting pulse distributor, and a switching and shorting pulse combiner. A gating signal is produced based on a carrier current and three modulating signals.
A shorting pulse is added to this signal when no top switches and no bottom switches are gated, causing the RMS currents to be equal in all legs. The same methods are utilized for each phase, however, switching variables are 120 degrees out of phase relative to one another, and the current pulses are shifted by a half-cycle with respect to output currents. If a triangular carrier is used with sinusoidal modulating signals, the CSI is said to be utilizing synchronized-pulse-width-modulation (SPWM). If full over-modulation is used in conjunction with SPWM the inverter is said to be in square-wave operation.
The second CSI modulation category, SHE is also similar to its VSI counterpart. Utilizing the gating signals developed for a VSI and a set of synchronizing sinusoidal current signals, results in symmetrically distributed shorting pulses and, therefore, symmetrical gating patterns. This allows any arbitrary number of harmonics to be eliminated. It also allows control of the fundamental line current through the proper selection of primary switching angles. Optimal switching patterns must have quarter-wave and half-wave symmetry, as well as symmetry about 30 degrees and 150 degrees. Switching patterns are never allowed between 60 degrees and 120 degrees. The current ripple can be further reduced with the use of larger output capacitors, or by increasing the number of switching pulses.
The third category, space-vector-based modulation, generates PWM load line currents that equal load line currents, on average. Valid switching states and time selections are made digitally based on space vector transformation. Modulating signals are represented as a complex vector using a transformation equation. For balanced three-phase sinusoidal signals, this vector becomes a fixed module, which rotates at a frequency, ω. These space vectors are then used to approximate the modulating signal. If the signal is between arbitrary vectors, the vectors are combined with the zero vectors I7, I8, or I9.[15] The following equations are used to ensure that the generated currents and the current vectors are on average equivalent.

Multilevel inverters

FIGURE 10: Three-Level Neutral-Clamped Inverter
A relatively new class called multilevel inverters has gained widespread interest. Normal operation of CSIs and VSIs can be classified as two-level inverters because the power switches connect to either the positive or the negative DC bus. If more than two voltage levels were available to the inverter output terminals, the AC output could better approximate a sine wave. For this reason multilevel inverters, although more complex and costly, offer higher performance. A three-level neutral-clamped inverter is shown in Figure 10.
Control methods for a three-level inverter only allow two switches of the four switches in each leg to simultaneously change conduction states. This allows smooth commutation and avoids shoot through by only selecting valid states. It may also be noted that since the DC bus voltage is shared by at least two power valves, their voltage ratings can be less than a two-level counterpart.
Carrier-based and space-vector modulation techniques are used for multilevel topologies. The methods for these techniques follow those of classic inverters, but with added complexity. Space-vector modulation offers a greater number of fixed voltage vectors to be used in approximating the modulation signal, and therefore allows more effective space vector PWM strategies to be accomplished at the cost of more elaborate algorithms. Due to added complexity and number of semiconductor devices, multilevel inverters are currently more suitable for high-power high-voltage applications. This technology reduces the harmonics hence improves overall efficiency of the scheme.


     
          
                                                  🔁Mars Exploration Rovers🔃 
                  Opportunity Self-Portrait


Technologies of Broad Benefit:
Power

Power technologies provide more efficient and increased electricity to the spacecraft and its subsystems.

Evolving the flat, square solar panel from the Sojourner rover on the 1996 Mars Pathfinder mission into more powerful solar "wings"

                                             
                                        Over a dozen engineers and technicians in white and blue cleanroom smocks and bonnets look over the Spirit rover in JPL's large spacecraft assembly room. The solar panels are down so the team can delicately place solar cells on the rover 'wings.' These cells absorb sunlight, which powers the rover.
This image is a bird's eye view of a group of engineers and technicians who are working on placing solar cells on the Spirit rover's 'wings' in JPL's spacecraft assembly facility.
Download full image ›
Like Mars Pathfinder, the Mars Exploration Rovers are solar-powered. Energy is absorbed through solar arrays on panels that sit atop what look like the rover's "wings." They were designed to maximize the area of solar cells that collect the sun's energy.

Another innovation for the rovers is the addition of Triple Junction Gallium Arsenides. These three-layered solar cells made their first trip to Mars aboard the twin rovers. Used on NASA's Deep Space 1 mission, these cells are able to absorb more sunlight than the single cell versions sent on Sojourner. The solar cells are stacked in three layers on the rover's solar arrays and, because they absorb more sunlight, can supply more power to the rover's re-chargeable lithium batteries.

The Sojourner rover on the Pathfinder mission carried one 40-amp-hour lithium battery. The Mars Exploration Rovers carry two 8-amp-hour lithium batteries. During the rovers' prime missions, their solar arrays were able to produce about 900 watt-hours of energy per martian day, or sol. Well into the extended mission, efforts to drive Spirit and Opportunity strategically through and toward solar-rich areas is providing up to 410 watt-hours per martian sol.

Using solar power limits the places on Mars that landed rover missions can explore. They are restricted to landing and traveling around the equatorial region where they can get enough sunlight to re-energize their batteries. For future missions, NASA is considering alternate power sources to increase the area on Mars that might be studied, opening up the whole planet to exploration.
      


  
    

              Mars Exploration Rovers:
                           Spacecraft


What is the spacecraft?

                                         Spacecraft

                                                              Spacecraft
Download full image ›
The spacecraft is separate from the launch vehicle. It is the protective "spaceship" that enables the precise cargo (that is, the rover!) to travel between Earth and Mars once the launch vehicle has already projected it outside of Earth´s atmosphere and gravity pull.
The spacecraft includes the mechanical units that safely carry and maneuver the rover as it enters the Martian atmosphere and lands on Mars. Once on the surface, like a semi-truck carrying a racecar cross-country, it lowers a ramp to let the rover drive out.



 Entry, Descent, & Landing System:
Configuration for entry into the Martian atmosphere. Includes the aeroshell (the heatshield and backshell), the parachute, the airbags, and a lander structure.

Rover:
A wheeled vehicle with science instruments for discoveries on the Martian surface.
The spacecraft design for the Mars Exploration Rover mission is largely based on the successful Mars Pathfinder system for entry, descent, and landing. The rover design is based on the Athena Rover on the previously cancelled Mars 2001 lander mission.


How much does the spacecraft weigh?

Each system has a total launch mass of 1,063 kilograms (2,343 pounds). The mass of each primary part of the spacecraft is as follows:

   Allocated Mass in kg
(lbs)
Cumulative Mass in kg
(lbs)
Rover
185 kg
(408 lbs)
185 kg
(408 lbs)
Lander
348 kg
(767 lbs)
533 kg
(1,175 lbs)
Backshell / Parachute
209 kg
(742 lbs)
742 kg
(1,636 lbs)
Heat Shield
78 kg
(172 lbs)
820 kg
(1,808 lbs)
Cruise Stage
193 kg
(425 lbs)
1,013 kg
(2,233 lbs)
Propellant
50 kg
(110 lbs)
1,063 kg
(2,343 lbs)




         
                                  Artist's concept of the Mars Helicopter



                     How Drone Solutions Are Powering the Future of Solar Energy

We explore how resources such as solar energy can benefit from the technology of the future, specifically drone based solutions. We’ll explore how each stage of setting up solar power stations can use drone solutions to bring light to the future.

The use of drone solutions can prevent and pre-emptively predict the spread of degradation of solar panels and determine which panels need cleaning or repair. Drone solutions that detect hotspots and drop in energy outputs are able not only to increase the PV plant efficiency but also reduce the cost of its maintenance since they practically require any human participation in work. Drone solutions in this way help teams pinpoint problem areas and fast-track repairs to keep solar operation running at peak efficiency and reduce the number of manpower per MW needed for the upkeep in performance.

                         Unmanned Drones Degradation and Hotspots Detection of Solar Panels

Future of drones : beyond the horizon

Drone solutions have unimaginable possibilities to revolutionise energy industries by automating the way we plan, build and maintain sustainable energy sources, and completing tasks faster, cheaper, and more safely than humans ever could, or even performing functions that could never have been dreamed of in the first place without drone technology.
Solar energy’s overall share of global power generation remains low, but is about to witness a major increase with the onset of drones into the sector. With technologies like thermal imagery and 3D modelling, drone solutions are set to cause a major paradigm shift as a growing number of companies are accommodating drone based outputs into their workflows, as the world shifts to new resources and companies are built to harness the resources so will solutions that enable this shift, stay tuned to understand in even more depth how drone solutions can be used in each stage of the project lifecycle.



                                               🔝 STAR LIGHT ENERGY⦽

Starlight (interstellar probe) Project Starlight is a research project of the University of California, Santa Barbara to develop a fleet of laser beam-propelled spacecraft and sending them to a star neighboring the Solar System, potentially Alpha Centauri. The project aims to send organisms on board the spacecraft

Starlight aims to accelerate the spacecrafts with powerful lasers, a method the project refers to as DEEP-IN (Directed Energy Propulsion for Interstellar Exploration), thus allowing them to reach stars near the Solar System in a matter of years, in contrast to traditional propulsion methods which will require thousands of years. Each spacecraft will be the size of a DVD disc and will be powered by plutonium. They will fly at one-fifth of the speed of light, and in the case of Alpha Centauri, it will arrive after traveling more than twenty years from Earth. Starlight is a program of the Experimental Cosmology Group of University of California, Santa Barbara (UCSB), and has received funding from NASA. In 2015, the NASA Innovative Advanced Concepts (NIAC) selected DEEP-IN as a phase-1 project.


One goal of Starlight is to send terrestrial organisms along with the spacecraft, and observe how the interstellar environment and extreme acceleration affects them. This effort is known as Terrestrial Biomes in Space, and the lead candidate is Caenorhabditis elegans, a minuscule nematode. The organism will spend most of the voyage in a frozen state, and once the spacecraft approaches its target they will be thawed by heat from the onboard plutonium. Following their revival, the organisms will be monitored by various sensors, and the data they produce will be sent back to Earth. C. elegans have been used extensively in biological research as a model organism, owing to the fact that the worm has one of the least number of cells for an animal possessing a nervous system. A backup option for C. elegans are tardigrades, micro-animals that are known for their resilience to various conditions lethal to other animals, such as the vacuum environment of space and strong doses of ionizing radiation.

Planetary protection

NASA's funding does not cover the Terrestrial Biome in Space portion of Starlight, as the experiment may potentially contaminate exoplanets.

Interstellar travel is crewed or uncrewed travel between stars or planetary systems. Interstellar travel would be much more difficult than interplanetary spaceflight. Whereas the distances between the planets in the Solar System are less than 30 astronomical units (AU), the distances between stars are typically hundreds of thousands of AU, and usually expressed in light-years. Because of the vastness of those distances, practical interstellar travel based on known physics would need to occur at a high percentage of the speed of light, allowing for significant travel times, at least decades to perhaps millennia or longer.
The speeds required for interstellar travel in a human lifetime far exceed what current methods of spacecraft propulsion can provide. Even with a hypothetically perfectly efficient propulsion system, the kinetic energy corresponding to those speeds is enormous by today's standards of energy development. Moreover, collisions by the spacecraft with cosmic dust and gas can produce very dangerous effects both to passengers and the spacecraft itself.
A number of strategies have been proposed to deal with these problems, ranging from giant arks that would carry entire societies and ecosystems, to microscopic space probes. Many different spacecraft propulsion systems have been proposed to give spacecraft the required speeds, including nuclear propulsion, beam-powered propulsion, and methods based on speculative physics.
For both crewed and uncrewed interstellar travel, considerable technological and economic challenges need to be met. Even the most optimistic views about interstellar travel see it as only being feasible decades from now. However, in spite of the challenges, if or when interstellar travel is realized, a wide range of scientific benefits is expected.
Most interstellar travel concepts require a developed space logistics system capable of moving millions of tonnes to a construction / operating location, and most would require gigawatt-scale power for construction or power (such as Star Wisp or Light Sail type concepts). Such a system could grow organically if space-based solar power became a significant component of Earth's energy mix. Consumer demand for a multi-terawatt system would automatically create the necessary multi-million ton/year logistical system
                             Mars Ice Home concept


Theoretical concepts

Faster-than-light travel

Artist's depiction of a hypothetical Wormhole Induction Propelled Spacecraft, based loosely on the 1994 "warp drive" paper of Miguel Alcubierre.
 
Scientists and authors have postulated a number of ways by which it might be possible to surpass the speed of light, but even the most serious-minded of these are highly speculative.
It is also debatable whether faster-than-light travel is physically possible, in part because of causality concerns: travel faster than light may, under certain conditions, permit travel backwards in time within the context of special relativity. Proposed mechanisms for faster-than-light travel within the theory of general relativity require the existence of exotic matter and it is not known if this could be produced in sufficient quantity.
Alcubierre drive
In physics, the Alcubierre drive is based on an argument, within the framework of general relativity and without the introduction of wormholes, that it is possible to modify spacetime in a way that allows a spaceship to travel with an arbitrarily large speed by a local expansion of spacetime behind the spaceship and an opposite contraction in front of it. Nevertheless, this concept would require the spaceship to incorporate a region of exotic matter, or hypothetical concept of negative mass.
Artificial black hole
A theoretical idea for enabling interstellar travel is by propelling a starship by creating an artificial black hole and using a parabolic reflector to reflect its EINSTEIN Gravitation 5 dimension . Although beyond current technological capabilities, a black hole star ship offers some advantages compared to other possible methods. Getting the black hole  ( Long Transverse ) to act as a power source and engine also requires a way to convert the Hawking radiation into energy and thrust. One potential method involves placing the hole at the focal point of a parabolic reflector attached to the ship, creating forward thrust. A slightly easier, but less efficient method would involve simply absorbing all the gamma radiation heading towards the fore of the ship to push it on wards, and let the rest shoot out the back.


_________________________________________________________________________________

                                          Hasil gambar untuk spaceship with starlight energy

         Gen. Mac Tech making 2 directions into 1 space fighter aircraft energy technique
_________________________________________________________________________________


Sabtu, 07 Maret 2020

JESI ISE Loving 2 in 1 to prediction accuracy of e- SAN ( Source Account Numerical ) press and pull, energy and energy transfer, computer processes Read and write and copy paste, are numerical mathematical values that are derived (set) and integral (reset), even though in practice numerical values are lost in the area of certain minimum and maximum values due to neutral harmonization but we can penetrate harmonization values close to harmonics 1, 2, 3, 4 due to the limitations of electronic sensor aids in the 20th and 21st century if and only if neutral harmonics can be counted will the results be void like space that is empty and limited by us humans.




All life processes always begin and are initiated by the concept of attraction (Push and Pull), some of which are:
1. Love which is the attraction between men and women
     towards the stability of body and soul.
2. energy conservation where the numerical value of energy moves in a manner
    push - pull goes to neutral ie 0 (null or void) or
    the harmonization process also leads to core stability although it can
    into a different form but the taste is the same.
3. read and write memory and copy paste which is
    Data collection (Push) also processes data into results
    Legitimate information is a mechanism (Pull).
4. Push (collect) and delete (Pull).
5. electronic circuits which are also systems of attraction
    where energy will turn into motion, light, indicators
    and heat is also a magnet until such a function is stable
    Batteries, batteries, adapters (there are harmonizations 1, 2, 3, 4 are called
    ripple), electronic components and their circuit pathways.
6. Space is also a process of attraction of various kinds
    kinds of harmonizing forces and energies that move in space
    and time so that they experience emptiness or vacuum,
    or we can call space is a neutral zone (0)


                                          Gen. Mac Tech to operation and change operation
                                                                        
                                                                             1

                             Hasil gambar untuk change operating system


                                                      (     AMNIMARJESLO   SAN    )











 

                                      2

        ⇉    Push–pull Input / Output  ⇇ 

                     Hasil gambar untuk push pull electronic concept

A push–pull in electronic is a concept of attraction in electronics is a special concept because all electronic components function due to the attractive relationship between positive and negative electrons, as well as the attraction of photons in the electronic components of light. Thus humans in life there is also the process of attraction that is love. 

electronic circuit that uses a pair of active devices that alternately supply current to, or absorb current from, a connected load. Push–pull outputs are present in TTL and CMOS digital logic circuits and in some types of amplifiers, and are usually realized as a complementary pair of transistors, one dissipating or sinking current from the load to ground or a negative power supply, and the other supplying or sourcing current to the load from a positive power supply ,
Push–pull circuits are widely used in many amplifier output stages . 

Push–pull transistor output stages :  

Categories include:

Transformer-output transistor power amplifiers

It is now very rare to use output transformers with transistor amplifiers, although such amplifiers offer the best opportunity for matching the output devices (with only PNP or only NPN devices required).

Totem pole push–pull output stages

Two matched transistors of the same polarity can be arranged to supply opposite halves of each cycle without the need for an output transformer, although in doing so the driver circuit often is asymmetric and one transistor will be used in a common-emitter configuration while the other is used as an emitter follower. This arrangement is less used today than during the 1970's; it can be implemented with few transistors (not so important today) but is relatively difficult to balance and to keep a low distortion.

Symmetrical push–pull

Each half of the output pair "mirror" the other, in that an NPN (or N-Channel FET) device in one half will be matched by a PNP (or P-Channel FET) in the other. This type of arrangement tends to give lower distortion than quasi-symmetric stages because even harmonics are cancelled more effectively with greater symmetry.

Quasi-symmetrical push–pull

In the past when good quality PNP complements for high power NPN silicon transistors were limited, a workaround was to use identical NPN output devices, but fed from complementary PNP and NPN driver circuits in such a way that the combination was close to being symmetrical (but never as good as having symmetry throughout). Distortion due to mismatched gain on each half of the cycle could be a significant problem.

Super-symmetric output stages

Employing some duplication in the whole driver circuit, to allow symmetrical drive circuits can improve matching further, although driver asymmetry is a small fraction of the distortion generating process. Using a bridge-tied load arrangement allows a much greater degree of matching between positive and negative halves, compensating for the inevitable small differences between NPN and PNP devices.

Square-law push–pull

The output devices, usually MOSFETs or vacuum tubes, are configured so that their square-law transfer characteristics (that generate second-harmonic distortion if used in a single-ended circuit) cancel distortion to a large extent. That is, as one transistor's gate-source voltage increases, the drive to the other device is reduced by the same amount and the drain (or plate) current change in the second device approximately corrects for the non-linearity in the increase of the first.

Push–pull tube (valve) output stages

Vacuum tubes (valves) are not available in complementary types (as are pnp/npn transistors), so the tube push–pull amplifier has a pair of identical output tubes or groups of tubes with the control grids driven in antiphase. These tubes drive current through the two halves of the primary winding of a center-tapped output transformer. Signal currents add, while the distortion signals due to the non-linear characteristic curves of the tubes subtract. These amplifiers were first designed long before the development of solid-state electronic devices; they are still in use by both audiophiles and musicians who consider them to sound better.

Vacuum tube push–pull amplifiers usually use an output transformer, although Output-transformerless (OTL) tube stages exist (such as the SEPP/SRPP and the White Cathode Follower below). The phase-splitter stage is usually another vacuum tube but a transformer with a center-tapped secondary winding was occasionally used in some designs. Because these are essentially square-law devices, the comments regarding distortion cancellation mentioned above apply to most push–pull tube designs when operated in class A (i.e. neither device is driven to its non-conducting state).

A Single Ended Push–Pull (SEPP, SRPP or mu-follower) output stage, originally called the Series-Balanced amplifier (US patent 2,310,342, Feb 1943). is similar to a totem-pole arrangement for transistors in that two devices are in series between the power supply rails, but the input drive goes only to one of the devices, the bottom one of the pair; hence the (seemingly contradictory) Single-Ended description. The output is taken from the cathode of the top (not directly driven) device, which acts part way between a constant current source and a cathode follower but receiving some drive from the plate (anode) circuit of the bottom device. The drive to each tube therefore might not be equal, but the circuit tends to keep the current through the bottom device somewhat constant throughout the signal, increasing the power gain and reducing distortion compared with a true single-tube single-ended output stage.

The White Cathode Follower (Patent 2,358,428, Sep 1944 by E. L. C. White) is similar to the SEPP design above, but the signal input is to the top tube, acting as a cathode follower, but one where the bottom tube (in common cathode configuration) if fed (usually via a step-up transformer) from the current in the plate (anode) of the top device. It essentially reverses the roles of the two devices in SEPP. The bottom tube acts part way between a constant current sink and an equal partner in the push–pull workload. Again, the drive to each tube therefore might not be equal.
Transistor versions of the SEPP and White follower do exist, but are rare.

Ultra-linear push–pull

A so-called ultra-linear push–pull amplifier uses either pentodes or tetrodes with their screen grid fed from a percentage of the primary voltage on the output transformer. This gives efficiency and distortion that is a good compromise between triode (or triode-strapped) power amplifier circuits and conventional pentode or tetrode output circuits where the screen is fed from a relatively constant voltage source.

Push–pull converter

A push–pull converter is a type of DC-to-DC converter, a switching converter that uses a transformer to change the voltage of a DC power supply. The distinguishing feature of a push-pull converter is that the transformer primary is supplied with current from the input line by pairs of transistors in a symmetrical push-pull circuit. The transistors are alternately switched on and off, periodically reversing the current in the transformer. Therefore, current is drawn from the line during both halves of the switching cycle. This contrasts with buck-boost converters, in which the input current is supplied by a single transistor which is switched on and off, so current is only drawn from the line during half the switching cycle. During the other half the output power is supplied by energy stored in inductors or capacitors in the power supply. Push–pull converters have steadier input current, create less noise on the input line, and are more efficient in higher power applications.

 

                     
                                      

       
Push-pull converter (+12V → ±18V; 50W) as potted module. ① transformer; ② and ③ electrolytic capacitors vertical and horizontal mounted; ④ discrete circuit board in through-hole technology . 

Circuit operation

Full-bridge converter circuit diagram
Conceptual schematic of a full-bridge converter. This is not a center tapped or split primary push-pull converter.
Top: Simple inverter circuit shown with an electromechanical switch
and automatic equivalent
auto-switching device implemented with two transistors and split winding auto-transformer in place of the mechanical switch. 

The term push–pull is sometimes used to generally refer to any converter with bidirectional excitation of the transformer. For example, in a full-bridge converter, the switches (connected as an H-bridge) alternate the voltage across the supply side of the transformer, causing the transformer to function as it would for AC power and produce a voltage on its output side. However, push–pull more commonly refers to a two-switch topology with a split primary winding.
In any case, the output is then rectified and sent to the load. Capacitors are often included at the output to filter the switching noise.
In practice, it is necessary to allow a small interval between powering the transformer one way and powering it the other: the “switches” are usually pairs of transistors (or similar devices), and were the two transistors in the pair to switch simultaneously there would be a risk of shorting out the power supply. Hence, a small wait is needed to avoid this problem. This wait time is called "Dead Time" and is necessary to avoid transistor shoot-through.

Transistors

N-type and P-type power transistors can be used. Power MOSFETs are often chosen for this role due to their high current switching capability and their inherently low ON resistance. The gates or bases of the power transistors are tied via a resistor to one of the supply voltages. A P-type transistor is used to pull up the N-type power transistor gate (common source) and an N-type transistor is used to pull down the P-type power transistor gate.

Alternatively, all power transistors can be N-type, which offer around three times the gain of their P-type equivalents. In this alternative the N-type transistor used in place of the P-type has to be driven in this way: The voltage is amplified by one P-type transistor and one N-type transistor in common base configuration to rail-to-rail amplitude. Then the power transistor is driven in common drain configuration to amplify the current.

In high frequency applications both transistors are driven with common source.
The operation of the circuit means that both transistors are actually pushing, and the pulling is done by a low pass filter in general, and by a center tap of the transformer in the converter application. But because the transistors push in an alternating fashion, the device is called a push-pull converter.

Timing

If both transistors are in their on state, a short circuit results. On the other hand, if both transistors are in their off state, high voltage peaks appear due to back EMF.
If the driver for the transistors is powerful and fast enough, the back EMF has no time to charge the capacity of the windings and of the body-diode of the MOSFETs to high voltages.
If a microcontroller is used, it can be used to measure the peak voltage and digitally adjust the timing for the transistors, so that the peak only just appears. This is especially useful when the transistors are starting from cold with no peaks, and are in their boot phase.

The cycle starts with no voltage and no current. Then one transistor turns on, a constant voltage is applied to the primary, current increases linearly, and a constant voltage is induced in the secondary. After some time T the transistor is turned off, the parasitic capacities of the transistors and the transformer and the inductance of the transformer form an LC circuit which swings to the opposite polarity. Then the other transistor turns on. For the same time T charge flows back into the storage capacitor, then changes the direction automatically, and for another time T the charge flows in the transformer. Then again the first transistor turns on until the current is stopped. Then the cycle is finished, another cycle can start anytime later. The S-shaped current is needed to improve over the simpler converters and deal efficiently with remanence.

 

                                                                       3

⇉The conservation of energy on push - pull⇇

When you do a useful job with a force (a push or a pull), such as moving a car uphill, we say you're doing work, and that takes energy. ... So no energy is created or destroyed here: you're simply converting energy stored as fuel inside your body into potential energy stored by the car (because of its height).

Cars can pull off the same trick. Depending on which make and model you own, you probably know that it does so many kilometers or miles to the gallon; in other words, using a certain amount of energy-rich gasoline, it can transport you (and a moderate load) a certain distance down the road. What we have here are two examples of machines—the human body and the automobile—that obey one of the most important laws of physics: the conservation of energy. Written in its simplest form, it says that you can't create or destroy energy, but you can convert it from one form into another. Pretty much everything that happens in the universe obeys.

What is the conservation of energy?

The first thing we need to note is that the law of conservation of energy is completely different from energy conservation. Energy conservation means saving energy through such things as insulating your home or using public transportation; generally it saves you money and helps the planet. The conservation of energy has nothing to do with saving energy: it's all about where energy comes from and where it goes.
Write the law formally and it sounds like this:
In a closed system, the amount of energy is fixed. You can't create any more energy inside the system or destroy any of the energy that's already in there. But you can convert the energy you have from one form to another (and sometimes back again).
A "closed system" is a bit like a sealed box around whatever we're studying: no energy can leak into the box from the inside (or be introduced to the box from outside).
There are some even simpler, more familiar ways of stating the conservation of energy. "No pain, no gain" is a rough everyday equivalent: if you want something, you have to work for it. "There's no such thing as a free lunch" and "You don't get anything for free" are other examples.
                                      A house as an example of a closed energy system, with sunlight streaming into the house from outside.
Artwork: This house is an example of a closed system: the energy that's inside the red dotted line stays as it is or gets converted into other forms. We can't create any new energy inside the house out of nothing at all, and we can't make energy inside the house vanish without trace, though we can turn it into other forms. So what if one room of the house suddenly starts getting hotter? The heat energy making that happen must be coming from energy that's already inside the house in a different form (maybe it's wood in a fire that's being burned to release the chemical energy locked inside it). If that's not the case, we don't have a closed system: the "extra" heat must be coming into the house from outside (maybe strong sunlight streaming in through the window).

Examples of the conservation of energy

The conservation of energy (and the idea of a "closed system") sounds a bit abstract, but it becomes an awful lot clearer when we consider some real-life examples.

Driving a car

Fill a car up with gasoline and you have a closed system. All the energy you have at your disposal is locked inside the gas in your tank in chemical form. When the gas flows into your engine, it burns with oxygen in the air. The chemical energy in the gas is converted first into heat energy: the burning fuel makes hot expanding gas, which pushes the pistons in the engine cylinders. In this way, the heat is converted into mechanical energy. The pistons turn the crankshaft, gears, and driveshaft and—eventually—the car's wheels. As the wheels turn, they speed the vehicle along the road, giving it kinetic energy (energy of movement).
If a car were 100 percent efficient, all the chemical energy originally locked inside the gasoline would be converted into kinetic energy. Unfortunately, energy is wasted at each stage of this process. Some is lost to friction when metal parts rub and wear against one another and heat up; some energy is lost as sound (cars can be quite noisy—and sound is energy that has to come from somewhere) Not all the energy the car produces moves you down the road: quite a lot has to push against the air (so it's lost to air resistance or drag), while some will be used to power things like the headlights, air conditioning, and so on. Nevertheless, if you measure the energy you start with (in the gasoline) and calculate how much energy you finish with and lose on the way (everything from useful kinetic energy and useless energy lost to friction, sound, air resistance, and so on), you'll find the energy account always balances: the energy you start with is the energy you finish with.

                           Conservation of energy in action: the energy you put into a car, as fuel, is exactly balanced by the useful work the car does plus various energy losses.
Artwork: Like everything else, cars must obey the law of conservation of energy. They convert the energy in fuel into mechanical energy that moves you down the road, but waste quite a lot of energy in the process. If you put 100 units of energy into a car (in the form of fuel), only 15 units or so move you down the road. The rest is wasted as heat losses in the engine (74 percent); parasitic losses (6%, making electricity, for example, to light the headlamps); and drivetrain losses (5%, sending power to the wheels). The 15 useful units of energy are used to overcome drag (air resistance), friction (in the brakes), and rolling resistance (in the tires). Every bit of energy we put into a car has to go somewhere, so the energy outputs (74% + 6% + 5% + 15%) must always exactly add up to the original energy input (100%). Figures for city driving from Where the energy goes, fueleconomy.gov.
Now this only applies if your car is a "closed system." If you're driving along the straight and the road suddenly starts going downhill, you're going to be able to go much further than you'd be able to go otherwise. Does this violate the conservation of energy? No, because we're no longer dealing with a closed system. Your car is gaining kinetic energy from the gasoline in its tank, but it's also gaining kinetic energy because it's going downhill. This isn't a closed system so the conservation of energy doesn't apply anymore.

Boiling a kettle

                                                  A white plastic electric jug kettle.
Photo: An electric kettle like this converts electrical energy into heat energy. 

That's the reverse of the process that happens in the power plant that supplies your home, where electricity is produced using heat energy released by burning a fuel such as coal, oil, or gas.
Boil water with an electric kettle and you're seeing the conservation of energy at work again. Electrical energy drawn from the power outlet on your wall flows into the heating element in the base of your kettle. As the current flows through the element, the element rapidly heats up, so the electrical energy is converted into heat energy that gets passed to the cold water surrounding it. After a couple of minutes, the water boils and (if the power stays on) starts to turn to steam. How does the conservation of energy apply here? Most of the electrical energy that enters the kettle is converted into heat energy in the water, though some is used to provide latent heat of evaporation (the heat we need to give to liquids to turn them into gases such as steam). If you add up the total electrical energy "lost" by the electricity supply and the total energy gained by the water, you should find they're almost exactly the same. Why aren't they exactly equal? Simply because we don't have a closed system here. Some of the original energy is converted to sound and wasted (kettles can be quite noisy). Kettles also give off some heat to their surroundings—so that's also wasted energy.

Pushing a car uphill

In the everyday world, "work" is something you do to earn money; in physics, work has a different meaning. When you do a useful job with a force (a push or a pull), such as moving a car uphill, we say you're doing work, and that takes energy. If you push a car uphill, it has more potential energy at the top of the hill than it had at the bottom. Have you violated the conservation of energy by creating potential energy out of thin air?
No! To push the car, you have to do work against the force of gravity. Your body has to use energy to do work. Most of the energy your body uses is gained by the car as you push it uphill. The energy your body loses is pretty much equal to the work it does against gravity. And the energy the car gains is the same as the work done. So no energy is created or destroyed here: you're simply converting energy stored as fuel inside your body into potential energy stored by the car (because of its height).

Who discovered the conservation of energy?

How do we know the conservation of energy is true? First, it sounds sensible. If you put a heavy log on a fire it might burn for an hour. If you put a second log, roughly the same size, on the fire, it's reasonable to suppose you'll get twice as much heat or the fire will burn twice as long. By the same token, if five bananas can supply your body with an hour's energy, ten bananas should keep you running for two hours—although you might not enjoy guzzling them all at once! In other words, the energy in (the logs you add to the fire or the bananas you eat) is equal to the energy out (the heat you get by burning logs or the energy you make by eating bananas).
                      James Prescott Joule's experiment to demonstrate the mechanical equivalent of heat and the law of conservation of energy.
Photo: The Mechanical Equivalent of Heat: In James Prescott Joule's famous experiment, a falling weight (1) pulls on a rope that passes over a pulley (2). The rope spins an axle (3) that turns a paddle inside a sealed container of water (4). As the paddle spins, the water heats up. Joule proved that the heat energy gained by the water was exactly the same as the potential energy lost by the weight.
... the quantity of heat produced by the friction of bodies, whether solid or liquid, is always proportional to the quantity of force expended.
James Prescott Joule, The Mechanical Equivalent of Heat, 1845.
Reasonable guesswork doesn't quite cut the mustard in science. Really, we need to be sure that the energy we start with in a closed system is the same as the energy we end up with. So how do we know this? One of the first people to confirm the law of conservation of energy experimentally was English physicist James Prescott Joule (1818–1889), who used an ingenious bit of apparatus to find what he called "the mechanical equivalent of heat." He used a falling weight to drive a large paddle wheel sealed inside a container of water. He calculated the potential energy of the weight (the energy it had because of its height above Earth) and reasoned that, as the weight fell, it transferred pretty much all its energy to kinetic energy in the paddle wheel. As the paddle wheel turned, it stirred the water in the container and warmed it up by a small but significant amount. Now we know how much energy it takes to warm a certain mass of water by a certain number of degrees, so Joule was able to figure out how much energy the water had gained. To his delight, he found out that this figure exactly matched the energy lost by the falling weight. Joule's brilliant work on energy was recognized when the international scientific unit of energy (the joule) was named for him.
Joule built on earlier work by Anglo-American physicist Benjamin Thompson (1753–1814), also known as Count Rumford. While working in a Germany artillery factory, Rumford noted that cannon barrels got hot when they were being drilled out. He swiftly realized that the heat was not a magic property of the metal (as many people supposed) but came from the mechanical, frictional process of drilling: the more you drilled, the hotter the metal got. Rumford's simple calculations produced results that, according to Joule, were "not very widely different from that which I have deduced from my own experiments." That was a sign both men were on the right track.

Why perpetual motion machines never work

                             A perpetual motion machine that endlessly cycles water using a waterwheel and Archimedes screw
Back in the 19th century, charlatan inventors would pop up from time to time showing off miracle machines that seemed to be able to drive themselves forever. Inventions like this are called perpetual motion machines: they seem to be able to move forever without anyone adding any more energy. Often, machines like this were blatant tricks: the mechanisms were powered by a concealed assistant who sat in the shadows turning a hidden handle! Some of the machines sound plausible, but all of them unfortunately fall foul of the conservation of energy.
In one version of perpetual motion, illustrated here, water (1) tips down onto a waterwheel, turning it around (2). The turning wheel drives gears (3) that power an Archimedes screw, which lifts the water back up to the top, theoretically allowing the whole cycle to repeat itself forever. Although you might think energy is being recycled as the water moves around, it's also being lost all the time. The water at the top has potential energy and that can indeed drive a waterwheel as it falls. But some energy will be lost to friction as the wheel turns. More energy will be lost to friction in the gears and the screw. So, between them, the wheel, gears, and screw will not have the same amount of energy as the potential energy the water lost originally. That means the screw cannot lift as as much water back to the bath as fell down from it originally—so the machine will very quickly come to a stop.

What about the conservation of mass?

Nuclear reactions seem to create energy out of nothing breaking up or joining together atoms. Do they violate the conservation of energy? No!
Albert Einstein's famous equation E=mc2 shows that energy and mass are different forms of the same thing. Loosely speaking, you can convert a small amount of mass into a large amount of energy (as in a nuclear power plant, where large atoms split apart and give off energy in the process). Einstein's equation shows us we sometimes need to factor mass into the conservation of energy. In a nuclear reaction, we start off with one set of atoms (a certain amount of energy in the form of mass) and end up with a different set of atoms (a different amount of energy locked in their mass) plus energy that's released as heat. If we factor in the mass of the atoms before and after the reaction, plus the energy released in the process, we find the conservation of energy is satisfied exactly. Since mass is a form of energy, it's clear that we can't destroy mass or create it out of nothing in the same way that we can't create or destroy energy.




                                                                          4

⇉Push vs. pull: data movement for linked data structures⇇


As the performance gap between the CPU and main memory continues to grow, techniques to hide memory latency are essential to deliver a high performance computer system. Prefetching can often overlap memory latency with computation for array-based numeric applications. However, prefetching for pointer-intensive applications still remains a challenging problem. Prefetching linked data structures (LDS) is difficult because the address sequence of LDS traversal does not present the same arithmetic regularity as array-based applications and the data dependence of pointer dereferences can serialize the address generation process.

hardware/software mechanism to reduce memory access latencies for linked data structures. Instead of relying on the past address history to predict future accesses, we identify the load instructions that traverse the LDS, and execute them ahead of the actual computation. To overcome the serial nature of the LDS address generation, we attach a prefetch controller to each level of the memory hierarchy and push, rather than pull, data to the CPU. 


Stack (abstract data type)



In computer science, a stack is an abstract data type that serves as a collection of elements, with two principal operations:
  • push, which adds an element to the collection, and
  • pop, which removes the most recently added element that was not yet removed.
The order in which elements come off a stack gives rise to its alternative name, LIFO (last in, first out). Additionally, a peek operation may give access to the top without modifying the stack. The name "stack" for this type of structure comes from the analogy to a set of physical items stacked on top of each other, which makes it easy to take an item off the top of the stack, while getting to an item deeper in the stack may require taking off multiple other items first.

Considered as a linear data structure, or more abstractly a sequential collection, the push and pop operations occur only at one end of the structure, referred to as the top of the stack. This makes it possible to implement a stack as a singly linked list and a pointer to the top element. A stack may be implemented to have a bounded capacity. If the stack is full and does not contain enough space to accept an entity to be pushed, the stack is then considered to be in an overflow state. The pop operation removes an item from the top of the stack.
A stack is needed to implement depth-first search.

         
                                              
                 Similar to a stack of plates, adding or removing is only possible at the top

                           

          
               Simple representation of a stack runtime with push and pop operations.

 

            Hardware stack

A common use of stacks at the architecture level is as a means of allocating and accessing memory.

Basic architecture of a stack

A typical stack, storing local data and call information for nested procedure calls (not necessarily nested procedures). This stack grows downward from its origin. The stack pointer points to the current topmost datum on the stack. A push operation decrements the pointer and copies the data to the stack; a pop operation copies data from the stack and then increments the pointer. Each procedure called in the program stores procedure return information (in yellow) and local data (in other colors) by pushing them onto the stack. This type of stack implementation is extremely common, but it is vulnerable to buffer overflow attacks (see the text).
A typical stack is an area of computer memory with a fixed origin and a variable size. Initially the size of the stack is zero. A stack pointer, usually in the form of a hardware register, points to the most recently referenced location on the stack; when the stack has a size of zero, the stack pointer points to the origin of the stack.
The two operations applicable to all stacks are:
  • a push operation, in which a data item is placed at the location pointed to by the stack pointer, and the address in the stack pointer is adjusted by the size of the data item;
  • a pop or pull operation: a data item at the current location pointed to by the stack pointer is removed, and the stack pointer is adjusted by the size of the data item.
There are many variations on the basic principle of stack operations. Every stack has a fixed location in memory at which it begins. As data items are added to the stack, the stack pointer is displaced to indicate the current extent of the stack, which expands away from the origin.
Stack pointers may point to the origin of a stack or to a limited range of addresses either above or below the origin (depending on the direction in which the stack grows); however, the stack pointer cannot cross the origin of the stack. In other words, if the origin of the stack is at address 1000 and the stack grows downwards (towards addresses 999, 998, and so on), the stack pointer must never be incremented beyond 1000 (to 1001, 1002, etc.). If a pop operation on the stack causes the stack pointer to move past the origin of the stack, a stack underflow occurs. If a push operation causes the stack pointer to increment or decrement beyond the maximum extent of the stack, a stack overflow occurs.
Some environments that rely heavily on stacks may provide additional operations, for example:
  • Duplicate: the top item is popped, and then pushed again (twice), so that an additional copy of the former top item is now on top, with the original below it.
  • Peek: the topmost item is inspected (or returned), but the stack pointer and stack size does not change (meaning the item remains on the stack). This is also called top operation in many articles.
  • Swap or exchange: the two topmost items on the stack exchange places.
  • Rotate (or Roll): the n topmost items are moved on the stack in a rotating fashion. For example, if n=3, items 1, 2, and 3 on the stack are moved to positions 2, 3, and 1 on the stack, respectively. Many variants of this operation are possible, with the most common being called left rotate and right rotate.
Stacks are often visualized growing from the bottom up (like real-world stacks). They may also be visualized growing from left to right, so that "topmost" becomes "rightmost", or even growing from top to bottom. The important feature is that the bottom of the stack is in a fixed position. The illustration in this section is an example of a top-to-bottom growth visualization: the top (28) is the stack "bottom", since the stack "top" (9) is where items are pushed or popped from.
A right rotate will move the first element to the third position, the second to the first and the third to the second. Here are two equivalent visualizations of this process:
apple                         banana
banana    ===right rotate==>  cucumber
cucumber                      apple
cucumber                      apple
banana    ===left rotate==>   cucumber
apple                         banana
A stack is usually represented in computers by a block of memory cells, with the "bottom" at a fixed location, and the stack pointer holding the address of the current "top" cell in the stack. The top and bottom terminology are used irrespective of whether the stack actually grows towards lower memory addresses or towards higher memory addresses.
Pushing an item on to the stack adjusts the stack pointer by the size of the item (either decrementing or incrementing, depending on the direction in which the stack grows in memory), pointing it to the next cell, and copies the new top item to the stack area. Depending again on the exact implementation, at the end of a push operation, the stack pointer may point to the next unused location in the stack, or it may point to the topmost item in the stack. If the stack points to the current topmost item, the stack pointer will be updated before a new item is pushed onto the stack; if it points to the next available location in the stack, it will be updated after the new item is pushed onto the stack.
Popping the stack is simply the inverse of pushing. The topmost item in the stack is removed and the stack pointer is updated, in the opposite order of that used in the push operation.

Stack in main memory

Many CISC-type CPU designs, including the x86, Z80 and 6502, have a dedicated register for use as the call stack stack pointer with dedicated call, return, push, and pop instructions that implicitly update the dedicated register, thus increasing code density. Some CISC processors, like the PDP-11 and the 68000, also have special addressing modes for implementation of stacks, typically with a semi-dedicated stack pointer as well (such as A7 in the 68000). In contrast, most RISC CPU designs do not have dedicated stack instructions and therefore most if not all registers may be used as stack pointers as needed.

Stack in registers or dedicated memory

The x87 floating point architecture is an example of a set of registers organised as a stack where direct access to individual registers (relative the current top) is also possible. As with stack-based machines in general, having the top-of-stack as an implicit argument allows for a small machine code footprint with a good usage of bus bandwidth and code caches, but it also prevents some types of optimizations possible on processors permitting random access to the register file for all (two or three) operands. A stack structure also makes superscalar implementations with register renaming (for speculative execution) somewhat more complex to implement, although it is still feasible, as exemplified by modern x87 implementations.
Sun SPARC, AMD Am29000, and Intel i960 are all examples of architectures using register windows within a register-stack as another strategy to avoid the use of slow main memory for function arguments and return values.
There are also a number of small microprocessors that implements a stack directly in hardware and some microcontrollers have a fixed-depth stack that is not directly accessible. Examples are the PIC microcontrollers, the Computer Cowboys MuP21, the Harris RTX line, and the Novix NC4016. Many stack-based microprocessors were used to implement the programming language Forth at the microcode level. Stacks were also used as a basis of a number of mainframes and mini computers. Such machines were called stack machines, the most famous being the Burroughs B5000.

Applications of stacks

Expression evaluation and syntax parsing

Calculators employing reverse Polish notation use a stack structure to hold values. Expressions can be represented in prefix, postfix or infix notations and conversion from one form to another may be accomplished using a stack. Many compilers use a stack for parsing the syntax of expressions, program blocks etc. before translating into low level code. Most programming languages are context-free languages, allowing them to be parsed with stack based machines.

Backtracking

Another important application of stacks is backtracking. Consider a simple example of finding the correct path in a maze. There are a series of points, from the starting point to the destination. We start from one point. To reach the final destination, there are several paths. Suppose we choose a random path. After following a certain path, we realise that the path we have chosen is wrong. So we need to find a way by which we can return to the beginning of that path. This can be done with the use of stacks. With the help of stacks, we remember the point where we have reached. This is done by pushing that point into the stack. In case we end up on the wrong path, we can pop the last point from the stack and thus return to the last point and continue our quest to find the right path. This is called backtracking.
The prototypical example of a backtracking algorithm is depth-first search, which finds all vertices of a graph that can be reached from a specified starting vertex. Other applications of backtracking involve searching through spaces that represent potential solutions to an optimization problem. Branch and bound is a technique for performing such backtracking searches without exhaustively searching all of the potential solutions in such a space.

Compile time memory management

A number of programming languages are stack-oriented, meaning they define most basic operations (adding two numbers, printing a character) as taking their arguments from the stack, and placing any return values back on the stack. For example, PostScript has a return stack and an operand stack, and also has a graphics state stack and a dictionary stack. Many virtual machines are also stack-oriented, including the p-code machine and the Java Virtual Machine.
Almost all calling conventions‍—‌the ways in which subroutines receive their parameters and return results‍—‌use a special stack (the "call stack") to hold information about procedure/function calling and nesting in order to switch to the context of the called function and restore to the caller function when the calling finishes. The functions follow a runtime protocol between caller and callee to save arguments and return value on the stack. Stacks are an important way of supporting nested or recursive function calls. This type of stack is used implicitly by the compiler to support CALL and RETURN statements (or their equivalents) and is not manipulated directly by the programmer.
Some programming languages use the stack to store data that is local to a procedure. Space for local data items is allocated from the stack when the procedure is entered, and is deallocated when the procedure exits. The C programming language is typically implemented in this way. Using the same stack for both data and procedure calls has important security implications (see below) of which a programmer must be aware in order to avoid introducing serious security bugs into a program.

Efficient algorithms

Several algorithms use a stack (separate from the usual function call stack of most programming languages) as the principle data structure with which they organize their information. These include:

  • Graham scan, an algorithm for the convex hull of a two-dimensional system of points. A convex hull of a subset of the input is maintained in a stack, which is used to find and remove concavities in the boundary when a new point is added to the hull.
  • Part of the SMAWK algorithm for finding the row minima of a monotone matrix uses stacks in a similar way to Graham scan.
  • All nearest smaller values, the problem of finding, for each number in an array, the closest preceding number that is smaller than it. One algorithm for this problem uses a stack to maintain a collection of candidates for the nearest smaller value. For each position in the array, the stack is popped until a smaller value is found on its top, and then the value in the new position is pushed onto the stack.
  • The nearest-neighbor chain algorithm, a method for agglomerative hierarchical clustering based on maintaining a stack of clusters, each of which is the nearest neighbor of its predecessor on the stack. When this method finds a pair of clusters that are mutual nearest neighbors, they are popped and merged.

Security

Some computing environments use stacks in ways that may make them vulnerable to security breaches and attacks. Programmers working in such environments must take special care to avoid the pitfalls of these implementations.
For example, some programming languages use a common stack to store both data local to a called procedure and the linking information that allows the procedure to return to its caller. This means that the program moves data into and out of the same stack that contains critical return addresses for the procedure calls. If data is moved to the wrong location on the stack, or an oversized data item is moved to a stack location that is not large enough to contain it, return information for procedure calls may be corrupted, causing the program to fail.
Malicious parties may attempt a stack smashing attack that takes advantage of this type of implementation by providing oversized data input to a program that does not check the length of input. Such a program may copy the data in its entirety to a location on the stack, and in so doing it may change the return addresses for procedures that have called it. An attacker can experiment to find a specific type of data that can be provided to such a program such that the return address of the current procedure is reset to point to an area within the stack itself (and within the data provided by the attacker), which in turn contains instructions that carry out unauthorized operations.
This type of attack is a variation on the buffer overflow attack and is an extremely frequent source of security breaches in software, mainly because some of the most popular compilers use a shared stack for both data and procedure calls, and do not verify the length of data items. Frequently programmers do not write code to verify the size of data items, either, and when an oversized or undersized data item is copied to the stack, a security breach may occur.

                                        

                                                                     5

⇉  To operation and change operation in differential mathematics (SET) and integral (RESET)⇇

 

In calculus, Leibniz's rule for differentiation under the integral sign, named after Gottfried Leibniz, states that for an integral of the form
where , the derivative of this integral is expressible as
where the partial derivative indicates that inside the integral, only the variation of f(x, t) with x is considered in taking the derivative. Notice that if and are constants rather than functions of , we have a special case of Leibniz's rule:
Besides, if and , which is a common situation as well (for example, in the proof of Cauchy's repeated integration formula), we have:
Thus under certain conditions, one may interchange the integral and partial differential operators. This important result is particularly useful in the differentiation of integral transforms. An example of such is the moment generating function in probability theory, a variation of the Laplace transform, which can be differentiated to generate the moments of a random variable. Whether Leibniz's integral rule applies is essentially a question about the interchange of limits.

General form: Differentiation under the integral sign

Theorem. Let f(x, t) be a function such that both f(x, t) and its partial derivative fx(x, t) are continuous in t and x in some region of the (x, t)-plane, including a(x) ≤ tb(x), x0xx1. Also suppose that the functions a(x) and b(x) are both continuous and both have continuous derivatives for x0xx1. Then, for x0xx1,
This formula is the general form of the Leibniz integral rule and can be derived using the fundamental theorem of calculus. The (first) fundamental theorem of calculus is just the particular case of the above formula where a(x) = a, a constant, b(x) = x, and f(x, t) = f(t).
If both upper and lower limits are taken as constants, then the formula takes the shape of an operator equation:
where is the partial derivative with respect to and is the integral operator with respect to over a fixed interval. That is, it is related to the symmetry of second derivatives, but involving integrals as well as derivatives. This case is also known as the Leibniz integral rule.
The following three basic theorems on the interchange of limits are essentially equivalent:
  • the interchange of a derivative and an integral (differentiation under the integral sign; i.e., Leibniz integral rule);
  • the change of order of partial derivatives;
  • the change of order of integration (integration under the integral sign; i.e., Fubini's theorem).

Three-dimensional, time-dependent case

Figure 1: A vector field F(r, t) defined throughout space, and a surface Σ bounded by curve ∂Σ moving with velocity v over which the field is integrated.
A Leibniz integral rule for a two dimensional surface moving in three dimensional space is
where:
F(r, t) is a vector field at the spatial position r at time t,
Σ is a surface bounded by the closed curve ∂Σ,
dA is a vector element of the surface Σ,
ds is a vector element of the curve ∂Σ,
v is the velocity of movement of the region Σ,
∇⋅ is the vector divergence,
× is the vector cross product,
The double integrals are surface integrals over the surface Σ, and the line integral is over the bounding curve ∂Σ.

Higher dimensions

The Leibniz integral rule can be extended to multidimensional integrals. In two and three dimensions, this rule is better known from the field of fluid dynamics as the Reynolds transport theorem:
where is a scalar function, D(t) and ∂D(t) denote a time-varying connected region of R3 and its boundary, respectively, is the Eulerian velocity of the boundary (see Lagrangian and Eulerian coordinates) and dΣ = n dS is the unit normal component of the surface element.
The general statement of the Leibniz integral rule requires concepts from differential geometry, specifically differential forms, exterior derivatives, wedge products and interior products. With those tools, the Leibniz integral rule in n dimensions is[2]
where Ω(t) is a time-varying domain of integration, ω is a p-form, is the vector field of the velocity, denotes the interior product with , dxω is the exterior derivative of ω with respect to the space variables only and is the time derivative of ω.
However, all of these identities can be derived from a most general statement about Lie derivatives:
Here, the ambient manifold on which the differential form lives includes both space and time.
is the region of integration (a submanifold) at a given instant (it does not depend on , since its parametrization as a submanifold defines its position in time),
is the Lie derivative,
is the spacetime vector field obtained from adding the unitary vector field in the direction of time to the purely spatial vector field from the previous formulas (i.e, is the spacetime velocity of ),
is a diffeomorphism from the one-parameter group generated by the flow of , and
is the image of under such diffeomorphism.
Something remarkable about this form, is that it can account for the case when changes its shape and size over time, since such deformations are fully determined by .

Measure theory statement

Let be an open subset of , and be a measure space. Suppose satisfies the following conditions:
  1. is a Lebesgue-integrable function of for each .
  2. For almost all , the derivative exists for all .
  3. There is an integrable function such that for all and almost every .
Then by the dominated convergence theorem for all ,

Proofs

Proof of basic form

Let
By the definition of the derivative,
Substitute equation (1) into equation (2). The difference of two integrals equals the integral of the difference, and 1/h is a constant, so
Provided that the limit can be passed through the integral sign, we obtain
We claim that the passage of the limit under the integral sign is valid by the bounded convergence theorem (a corollary of the dominated convergence theorem). For each δ > 0, consider the difference quotient
For t fixed, the mean value theorem implies there exists z in the interval [x, x + δ] such that
Continuity of fx(x, t) and compactness of the domain together imply that fx(x, t) is bounded. The above application of the mean value theorem therefore gives a uniform (independent of δ) bound on . The difference quotients converge pointwise to the partial derivative fx by the assumption that the partial derivative exists.
The above argument shows that for every sequence {δn} → 0, the sequence is uniformly bounded and converges pointwise to fx. The bounded convergence theorem states that if a sequence of functions on a set of finite measure is uniformly bounded and converges pointwise, then passage of the limit under the integral is valid. In particular, the limit and integral may be exchanged for every sequence {δn} → 0. Therefore, the limit as δ → 0 may be passed through the integral sign.
For a simpler proof using Fubini's theorem, see the references.

Variable limits form

For a continuous real valued function g of one real variable, and real valued differentiable functions and of one real variable,
This follows from the chain rule and the First Fundamental Theorem of Calculus. Define
,
and
. (The lower limit just has to be some number in the domain of )
Then, can be written as a composition: . The Chain Rule then implies that
.
By the First Fundamental Theorem of Calculus, . Therefore, substituting this result above, we get the desired equation:
.
Note: This form can be particularly useful if the expression to be differentiated is of the form:
Because does not depend on the limits of integration, it may be move out from under the integral sign, and the above form may be used with the Product rule, i.e.

General form with variable limits

Set
where a and b are functions of α that exhibit increments Δa and Δb, respectively, when α is increased by Δα. Then,
A form of the mean value theorem, , where a < ξ < b, may be applied to the first and last integrals of the formula for Δφ above, resulting in
Divide by Δα and let Δα → 0. Notice ξ1a and ξ2b. We may pass the limit through the integral sign:
again by the bounded convergence theorem. This yields the general form of the Leibniz integral rule,

Alternative Proof of General Form with Variable Limits, using the Chain Rule

The general form of Leibniz's Integral Rule with variable limits can be derived as a consequence of the basic form of Leibniz's Integral Rule, the Multivariable Chain Rule, and the First Fundamental Theorem of Calculus. Suppose is defined in a rectangle in the plane, for and . Also, assume and the partial derivative are both continuous functions on this rectangle. Suppose are differentiable real valued functions defined on , with values in (i.e. for every ). Now, set
,   for and
and
,   for
Then, by properties of Definite Integrals, we can write
Since the functions are all differentiable (see the remark at the end of the proof), by the Multivariable Chain Rule, it follows that is differentiable, and its derivative is given by the formula:
 
Now, note that for every , and for every , we have that , because when taking the partial derivative with respect to of , we are keeping fixed in the expression ; thus the basic form of Leibniz's Integral Rule with constant limits of integration applies. Next, by the First Fundamental Theorem of Calculus, we have that ; because when taking the partial derivative with respect to of , the first variable is fixed, so the fundamental theorem can indeed be applied.
Substituting these results into the equation for above gives:
as desired.
There is a technical point in the proof above which is worth noting: applying the Chain Rule to requires that already be Differentiable. This is where we use our assumptions about . As mentioned above, the partial derivatives of are given by the formulas and . Since is continuous, its integral is also a continuous function, and since is also continuous, these two results show that both the partial derivatives of are continuous. Since continuity of partial derivatives implies differentiability of the function, is indeed differentiable.

Three-dimensional, time-dependent form

At time t the surface Σ in Figure 1 contains a set of points arranged about a centroid . The function can be written as
with independent of time. Variables are shifted to a new frame of reference attached to the moving surface, with origin at . For a rigidly translating surface, the limits of integration are then independent of time, so:
where the limits of integration confining the integral to the region Σ no longer are time dependent so differentiation passes through the integration to act on the integrand only:
with the velocity of motion of the surface defined by
This equation expresses the material derivative of the field, that is, the derivative with respect to a coordinate system attached to the moving surface. Having found the derivative, variables can be switched back to the original frame of reference. We notice that (see article on curl)
and that Stokes theorem equates the surface integral of the curl over Σ with a line integral over ∂Σ:
The sign of the line integral is based on the right-hand rule for the choice of direction of line element ds. To establish this sign, for example, suppose the field F points in the positive z-direction, and the surface Σ is a portion of the xy-plane with perimeter ∂Σ. We adopt the normal to Σ to be in the positive z-direction. Positive traversal of ∂Σ is then counterclockwise (right-hand rule with thumb along z-axis). Then the integral on the left-hand side determines a positive flux of F through Σ. Suppose Σ translates in the positive x-direction at velocity v. An element of the boundary of Σ parallel to the y-axis, say ds, sweeps out an area vt × ds in time t. If we integrate around the boundary ∂Σ in a counterclockwise sense, vt × ds points in the negative z-direction on the left side of ∂Σ (where ds points downward), and in the positive z-direction on the right side of ∂Σ (where ds points upward), which makes sense because Σ is moving to the right, adding area on the right and losing it on the left. On that basis, the flux of F is increasing on the right of ∂Σ and decreasing on the left. However, the dot product v × F • ds = −F × vds = −F • v × ds. Consequently, the sign of the line integral is taken as negative.
If v is a constant,
which is the quoted result. This proof does not consider the possibility of the surface deforming as it moves.

Alternative derivation

Lemma. One has:
Proof. From proof of the fundamental theorem of calculus,
and
Suppose a and b are constant, and that f(x) involves a parameter α which is constant in the integration but may vary to form different integrals. Assume that f(x, α) is a continuous function of x and α in the compact set {(x, α) : α0 ≤ α ≤ α1 and axb}, and that the partial derivative fα(x, α) exists and is continuous. If one defines:
then may be differentiated with respect to α by differentiating under the integral sign, i.e.,
By the Heine–Cantor theorem it is uniformly continuous in that set. In other words, for any ε > 0 there exists Δα such that for all values of x in [a, b],
On the other hand,
Hence φ(α) is a continuous function.
Similarly if exists and is continuous, then for all ε > 0 there exists Δα such that:
Therefore,
where
Now, ε → 0 as Δα → 0, so
This is the formula we set out to prove.
Now, suppose
where a and b are functions of α which take increments Δa and Δb, respectively, when α is increased by Δα. Then,
A form of the mean value theorem, where a < ξ < b, can be applied to the first and last integrals of the formula for Δφ above, resulting in
Dividing by Δα, letting Δα → 0, noticing ξ1a and ξ2b and using the above derivation for
yields
This is the general form of the Leibniz integral rule.

Examples

General examples

Example 1

Consider the function
The function under the integral sign is not continuous at the point (x, α) = (0, 0), and the function φ(α) has a discontinuity at α = 0 because φ(α) approaches ±π/2 as α → 0±.
If we differentiate φ(α) with respect to α under the integral sign, we get
which is, of course, true for all values of α except α = 0. This may be integrated (with respect to α) to find

Example 2

An example with variable limits:

Examples for evaluating a definite integral

Example 3

The principle of differentiating under the integral sign may sometimes be used to evaluate a definite integral. Consider:
Now,
As x varies from 0 to π, we have
Hence,
Therefore,
Integrating both sides with respect to α, we get:
C1 = 0 follows from evaluating φ(0):
To determine C2 in the same manner, we should need to substitute in a value of α greater than 1 in φ(α). This is somewhat inconvenient. Instead, we substitute α = 1/β, where |β| < 1. Then,
Therefore, C2 = 0.
The definition of φ(α) is now complete:
The foregoing discussion, of course, does not apply when α = ±1, since the conditions for differentiability are not met.

Example 4

First we calculate:
The limits of integration being independent of a, we have:
On the other hand:
Equating these two relations then yields
In a similar fashion, pursuing yields
Adding the two results then produces
which computes as desired.
This derivation may be generalized. Note that if we define
it can easily be shown that
Given I1, this integral reduction formula can be used to compute all of the values of In for n > 1.

Example 5

Here, we consider the integral
Differentiating under the integral with respect to α, we have
Therefore:
However, by definition, I(π/2) = 0, hence C = π2/8 and

Example 6

Here, we consider the integral
We introduce a new variable φ and rewrite the integral as
When φ = 1 this equals the original integral. However, this more general integral may be differentiated with respect to φ:
This is the line integral of over the unit circle. By Green's Theorem, it equals the double integral over the unit disk of which equals 0. This implies that f(φ) is constant. The constant may be determined by evaluating f at φ = 0:
Therefore, the original integral also equals 2π.

Other problems to solve

There are innumerable other integrals that can be solved using the technique of differentiation under the integral sign. For example, in each of the following cases, the original integral may be replaced by a similar integral having a new parameter α:
The first integral, the Dirichlet integral, is absolutely convergent for positive α but only conditionally convergent when α is 0. Therefore, differentiation under the integral sign is easy to justify when α > 0, but proving that the resulting formula remains valid when α is 0 requires some careful work.

Applications to series

The measure-theoretic version of differentiation under the integral sign also applies to summation (finite or infinite) by interpreting summation as counting measure. An example of an application is the fact that power series are differentiable in their radius of convergence.

In popular culture

Differentiation under the integral sign is mentioned in the late physicist Richard Feynman's best-selling memoir Surely You're Joking, Mr. Feynman! in the chapter "A Different Box of Tools". He describes learning it, while in high school, from an old text, Advanced Calculus (1926), by Frederick S. Woods (who was a professor of mathematics in the Massachusetts Institute of Technology). The technique was not often taught when Feynman later received his formal education in calculus, but using this technique, Feynman was able to solve otherwise difficult integration problems upon his arrival at graduate school at Princeton University:
One thing I never did learn was contour integration. I had learned to do integrals by various methods shown in a book that my high school physics teacher Mr. Bader had given me. One day he told me to stay after class. "Feynman," he said, "you talk too much and you make too much noise. I know why. You're bored. So I'm going to give you a book. You go up there in the back, in the corner, and study this book, and when you know everything that's in this book, you can talk again." So every physics class, I paid no attention to what was going on with Pascal's Law, or whatever they were doing. I was up in the back with this book: "Advanced Calculus", by Woods. Bader knew I had studied "Calculus for the Practical Man" a little bit, so he gave me the real works—it was for a junior or senior course in college. It had Fourier series, Bessel functions, determinants, elliptic functions—all kinds of wonderful stuff that I didn't know anything about. That book also showed how to differentiate parameters under the integral sign—it's a certain operation. It turns out that's not taught very much in the universities; they don't emphasize it. But I caught on how to use that method, and I used that one damn tool again and again. So because I was self-taught using that book, I had peculiar methods of doing integrals. The result was, when guys at MIT or Princeton had trouble doing a certain integral, it was because they couldn't do it with the standard methods they had learned in school. If it was contour integration, they would have found it; if it was a simple series expansion, they would have found it. Then I come along and try differentiating under the integral sign, and often it worked. .


                                                                      6

                                  ⇉   SPACE AND TIME on PUSH AND PULL ⇇

 

Both push and pull drive our galaxy's race through space


Discovery of the 'dipole repeller' confirms that both attraction and repulsion are at play in our extragalactic neighborhood   

Although we can't feel it, we're in constant motion: the earth spins on its axis at about 1,600 km/h; it orbits around the sun at about 100,000 km/h; the sun orbits our Milky Way galaxy at about 850,000 km/h; and the Milky Way galaxy and its companion galaxy Andromeda are moving with respect to the expanding universe at roughly 2 million km/h (630 km per second). But what is propelling the Milky Way's race through space?

Until now, scientists assumed that a dense region of the universe is pulling us toward it, in the same way that gravity made Newton's apple fall to earth. The initial "prime suspect" was called the Great Attractor, a region of a half dozen rich clusters of galaxies 150 million lightyears from the Milky Way. Soon after, attention was drawn to an area of more than two dozen rich clusters, called the Shapley Concentration, which sits 600 million lightyears beyond the Great Attractor.

By identifying the Dipole Repeller, the researchers were able to reconcile both the direction of the Milky Way's motion and its magnitude. They expect that future ultra-sensitive surveys at optical, near-infrared and radio wavelengths will directly identify the few galaxies expected to lie in this void, and directly confirm the void associated with the Dipole Repeller.   

Earth is a time space dimension, so it operates in a way that what fills the space then creates either distance or nearness, a longer time or a shorter time.

Congruence is an amazing word, when you allow the vibration of it to truly sink in. It describes a situation where in the moment (time), you are filling your cells and being with a message that all parts of you is agreeing on (space). If you understand this, then understand that everything that occurs as a response to you, offers you a mirror. For example, you’re a full yes to something, and then you present the idea to someone else and they’re a full no. Do you still remain a full yes? Or do you see it for what it is, you are in congrunece in this moment, and they are not? Do you use this as an opportunity to sharpen your articulation, so they are able to get a clear idea of just how congruent you are? Or do you slink away wounded, contracted and caved in? In response to a no do you strengthen your yes, so that you’re not just in agreement you are also amplified? 

If even a small part of you is out of alignment with the rest of you, then over time it will make itself known. That small part will convince you that it’s there to look after your best interest and keep you safe. You can consider it to be your gate keeper of sorts, it wants you to put your money where your mouth is. It will niggle, and niggle, and niggle, and niggle, in a quest for you to make the same decision over, and over, and over, and over. Eventually when enough of you is amplified you’ll  have pulled the keeper to your side, the gate opens and this becomes your new way of being. Instantly you feel the change because the world begins to respond very differently to you. That part of your life gets easier, just for a little while.

So pay attention to the areas where you’re broadcasting and amplifying a push pull type of energy. Sending mixed messages into the universe, also making it such that you’re not amplifed enough to become magnetic. You stengthen, then weaken, then weaken, then weaken, then strengthen. The energy is counter productive. Empower your stance, empower your decision and empower your actions.

 

  +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

                                Gen . Mac Tech in operation to change operation 

       
                          Hasil gambar untuk aircraft galaxy space

+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++