Sabtu, 07 Maret 2020

JESI ISE Loving 2 in 1 to prediction accuracy of e- SAN ( Source Account Numerical ) press and pull, energy and energy transfer, computer processes Read and write and copy paste, are numerical mathematical values that are derived (set) and integral (reset), even though in practice numerical values are lost in the area of certain minimum and maximum values due to neutral harmonization but we can penetrate harmonization values close to harmonics 1, 2, 3, 4 due to the limitations of electronic sensor aids in the 20th and 21st century if and only if neutral harmonics can be counted will the results be void like space that is empty and limited by us humans.




All life processes always begin and are initiated by the concept of attraction (Push and Pull), some of which are:
1. Love which is the attraction between men and women
     towards the stability of body and soul.
2. energy conservation where the numerical value of energy moves in a manner
    push - pull goes to neutral ie 0 (null or void) or
    the harmonization process also leads to core stability although it can
    into a different form but the taste is the same.
3. read and write memory and copy paste which is
    Data collection (Push) also processes data into results
    Legitimate information is a mechanism (Pull).
4. Push (collect) and delete (Pull).
5. electronic circuits which are also systems of attraction
    where energy will turn into motion, light, indicators
    and heat is also a magnet until such a function is stable
    Batteries, batteries, adapters (there are harmonizations 1, 2, 3, 4 are called
    ripple), electronic components and their circuit pathways.
6. Space is also a process of attraction of various kinds
    kinds of harmonizing forces and energies that move in space
    and time so that they experience emptiness or vacuum,
    or we can call space is a neutral zone (0)


                                          Gen. Mac Tech to operation and change operation
                                                                        
                                                                             1

                             Hasil gambar untuk change operating system


                                                      (     AMNIMARJESLO   SAN    )











 

                                      2

        ⇉    Push–pull Input / Output  ⇇ 

                     Hasil gambar untuk push pull electronic concept

A push–pull in electronic is a concept of attraction in electronics is a special concept because all electronic components function due to the attractive relationship between positive and negative electrons, as well as the attraction of photons in the electronic components of light. Thus humans in life there is also the process of attraction that is love. 

electronic circuit that uses a pair of active devices that alternately supply current to, or absorb current from, a connected load. Push–pull outputs are present in TTL and CMOS digital logic circuits and in some types of amplifiers, and are usually realized as a complementary pair of transistors, one dissipating or sinking current from the load to ground or a negative power supply, and the other supplying or sourcing current to the load from a positive power supply ,
Push–pull circuits are widely used in many amplifier output stages . 

Push–pull transistor output stages :  

Categories include:

Transformer-output transistor power amplifiers

It is now very rare to use output transformers with transistor amplifiers, although such amplifiers offer the best opportunity for matching the output devices (with only PNP or only NPN devices required).

Totem pole push–pull output stages

Two matched transistors of the same polarity can be arranged to supply opposite halves of each cycle without the need for an output transformer, although in doing so the driver circuit often is asymmetric and one transistor will be used in a common-emitter configuration while the other is used as an emitter follower. This arrangement is less used today than during the 1970's; it can be implemented with few transistors (not so important today) but is relatively difficult to balance and to keep a low distortion.

Symmetrical push–pull

Each half of the output pair "mirror" the other, in that an NPN (or N-Channel FET) device in one half will be matched by a PNP (or P-Channel FET) in the other. This type of arrangement tends to give lower distortion than quasi-symmetric stages because even harmonics are cancelled more effectively with greater symmetry.

Quasi-symmetrical push–pull

In the past when good quality PNP complements for high power NPN silicon transistors were limited, a workaround was to use identical NPN output devices, but fed from complementary PNP and NPN driver circuits in such a way that the combination was close to being symmetrical (but never as good as having symmetry throughout). Distortion due to mismatched gain on each half of the cycle could be a significant problem.

Super-symmetric output stages

Employing some duplication in the whole driver circuit, to allow symmetrical drive circuits can improve matching further, although driver asymmetry is a small fraction of the distortion generating process. Using a bridge-tied load arrangement allows a much greater degree of matching between positive and negative halves, compensating for the inevitable small differences between NPN and PNP devices.

Square-law push–pull

The output devices, usually MOSFETs or vacuum tubes, are configured so that their square-law transfer characteristics (that generate second-harmonic distortion if used in a single-ended circuit) cancel distortion to a large extent. That is, as one transistor's gate-source voltage increases, the drive to the other device is reduced by the same amount and the drain (or plate) current change in the second device approximately corrects for the non-linearity in the increase of the first.

Push–pull tube (valve) output stages

Vacuum tubes (valves) are not available in complementary types (as are pnp/npn transistors), so the tube push–pull amplifier has a pair of identical output tubes or groups of tubes with the control grids driven in antiphase. These tubes drive current through the two halves of the primary winding of a center-tapped output transformer. Signal currents add, while the distortion signals due to the non-linear characteristic curves of the tubes subtract. These amplifiers were first designed long before the development of solid-state electronic devices; they are still in use by both audiophiles and musicians who consider them to sound better.

Vacuum tube push–pull amplifiers usually use an output transformer, although Output-transformerless (OTL) tube stages exist (such as the SEPP/SRPP and the White Cathode Follower below). The phase-splitter stage is usually another vacuum tube but a transformer with a center-tapped secondary winding was occasionally used in some designs. Because these are essentially square-law devices, the comments regarding distortion cancellation mentioned above apply to most push–pull tube designs when operated in class A (i.e. neither device is driven to its non-conducting state).

A Single Ended Push–Pull (SEPP, SRPP or mu-follower) output stage, originally called the Series-Balanced amplifier (US patent 2,310,342, Feb 1943). is similar to a totem-pole arrangement for transistors in that two devices are in series between the power supply rails, but the input drive goes only to one of the devices, the bottom one of the pair; hence the (seemingly contradictory) Single-Ended description. The output is taken from the cathode of the top (not directly driven) device, which acts part way between a constant current source and a cathode follower but receiving some drive from the plate (anode) circuit of the bottom device. The drive to each tube therefore might not be equal, but the circuit tends to keep the current through the bottom device somewhat constant throughout the signal, increasing the power gain and reducing distortion compared with a true single-tube single-ended output stage.

The White Cathode Follower (Patent 2,358,428, Sep 1944 by E. L. C. White) is similar to the SEPP design above, but the signal input is to the top tube, acting as a cathode follower, but one where the bottom tube (in common cathode configuration) if fed (usually via a step-up transformer) from the current in the plate (anode) of the top device. It essentially reverses the roles of the two devices in SEPP. The bottom tube acts part way between a constant current sink and an equal partner in the push–pull workload. Again, the drive to each tube therefore might not be equal.
Transistor versions of the SEPP and White follower do exist, but are rare.

Ultra-linear push–pull

A so-called ultra-linear push–pull amplifier uses either pentodes or tetrodes with their screen grid fed from a percentage of the primary voltage on the output transformer. This gives efficiency and distortion that is a good compromise between triode (or triode-strapped) power amplifier circuits and conventional pentode or tetrode output circuits where the screen is fed from a relatively constant voltage source.

Push–pull converter

A push–pull converter is a type of DC-to-DC converter, a switching converter that uses a transformer to change the voltage of a DC power supply. The distinguishing feature of a push-pull converter is that the transformer primary is supplied with current from the input line by pairs of transistors in a symmetrical push-pull circuit. The transistors are alternately switched on and off, periodically reversing the current in the transformer. Therefore, current is drawn from the line during both halves of the switching cycle. This contrasts with buck-boost converters, in which the input current is supplied by a single transistor which is switched on and off, so current is only drawn from the line during half the switching cycle. During the other half the output power is supplied by energy stored in inductors or capacitors in the power supply. Push–pull converters have steadier input current, create less noise on the input line, and are more efficient in higher power applications.

 

                     
                                      

       
Push-pull converter (+12V → ±18V; 50W) as potted module. ① transformer; ② and ③ electrolytic capacitors vertical and horizontal mounted; ④ discrete circuit board in through-hole technology . 

Circuit operation

Full-bridge converter circuit diagram
Conceptual schematic of a full-bridge converter. This is not a center tapped or split primary push-pull converter.
Top: Simple inverter circuit shown with an electromechanical switch
and automatic equivalent
auto-switching device implemented with two transistors and split winding auto-transformer in place of the mechanical switch. 

The term push–pull is sometimes used to generally refer to any converter with bidirectional excitation of the transformer. For example, in a full-bridge converter, the switches (connected as an H-bridge) alternate the voltage across the supply side of the transformer, causing the transformer to function as it would for AC power and produce a voltage on its output side. However, push–pull more commonly refers to a two-switch topology with a split primary winding.
In any case, the output is then rectified and sent to the load. Capacitors are often included at the output to filter the switching noise.
In practice, it is necessary to allow a small interval between powering the transformer one way and powering it the other: the “switches” are usually pairs of transistors (or similar devices), and were the two transistors in the pair to switch simultaneously there would be a risk of shorting out the power supply. Hence, a small wait is needed to avoid this problem. This wait time is called "Dead Time" and is necessary to avoid transistor shoot-through.

Transistors

N-type and P-type power transistors can be used. Power MOSFETs are often chosen for this role due to their high current switching capability and their inherently low ON resistance. The gates or bases of the power transistors are tied via a resistor to one of the supply voltages. A P-type transistor is used to pull up the N-type power transistor gate (common source) and an N-type transistor is used to pull down the P-type power transistor gate.

Alternatively, all power transistors can be N-type, which offer around three times the gain of their P-type equivalents. In this alternative the N-type transistor used in place of the P-type has to be driven in this way: The voltage is amplified by one P-type transistor and one N-type transistor in common base configuration to rail-to-rail amplitude. Then the power transistor is driven in common drain configuration to amplify the current.

In high frequency applications both transistors are driven with common source.
The operation of the circuit means that both transistors are actually pushing, and the pulling is done by a low pass filter in general, and by a center tap of the transformer in the converter application. But because the transistors push in an alternating fashion, the device is called a push-pull converter.

Timing

If both transistors are in their on state, a short circuit results. On the other hand, if both transistors are in their off state, high voltage peaks appear due to back EMF.
If the driver for the transistors is powerful and fast enough, the back EMF has no time to charge the capacity of the windings and of the body-diode of the MOSFETs to high voltages.
If a microcontroller is used, it can be used to measure the peak voltage and digitally adjust the timing for the transistors, so that the peak only just appears. This is especially useful when the transistors are starting from cold with no peaks, and are in their boot phase.

The cycle starts with no voltage and no current. Then one transistor turns on, a constant voltage is applied to the primary, current increases linearly, and a constant voltage is induced in the secondary. After some time T the transistor is turned off, the parasitic capacities of the transistors and the transformer and the inductance of the transformer form an LC circuit which swings to the opposite polarity. Then the other transistor turns on. For the same time T charge flows back into the storage capacitor, then changes the direction automatically, and for another time T the charge flows in the transformer. Then again the first transistor turns on until the current is stopped. Then the cycle is finished, another cycle can start anytime later. The S-shaped current is needed to improve over the simpler converters and deal efficiently with remanence.

 

                                                                       3

⇉The conservation of energy on push - pull⇇

When you do a useful job with a force (a push or a pull), such as moving a car uphill, we say you're doing work, and that takes energy. ... So no energy is created or destroyed here: you're simply converting energy stored as fuel inside your body into potential energy stored by the car (because of its height).

Cars can pull off the same trick. Depending on which make and model you own, you probably know that it does so many kilometers or miles to the gallon; in other words, using a certain amount of energy-rich gasoline, it can transport you (and a moderate load) a certain distance down the road. What we have here are two examples of machines—the human body and the automobile—that obey one of the most important laws of physics: the conservation of energy. Written in its simplest form, it says that you can't create or destroy energy, but you can convert it from one form into another. Pretty much everything that happens in the universe obeys.

What is the conservation of energy?

The first thing we need to note is that the law of conservation of energy is completely different from energy conservation. Energy conservation means saving energy through such things as insulating your home or using public transportation; generally it saves you money and helps the planet. The conservation of energy has nothing to do with saving energy: it's all about where energy comes from and where it goes.
Write the law formally and it sounds like this:
In a closed system, the amount of energy is fixed. You can't create any more energy inside the system or destroy any of the energy that's already in there. But you can convert the energy you have from one form to another (and sometimes back again).
A "closed system" is a bit like a sealed box around whatever we're studying: no energy can leak into the box from the inside (or be introduced to the box from outside).
There are some even simpler, more familiar ways of stating the conservation of energy. "No pain, no gain" is a rough everyday equivalent: if you want something, you have to work for it. "There's no such thing as a free lunch" and "You don't get anything for free" are other examples.
                                      A house as an example of a closed energy system, with sunlight streaming into the house from outside.
Artwork: This house is an example of a closed system: the energy that's inside the red dotted line stays as it is or gets converted into other forms. We can't create any new energy inside the house out of nothing at all, and we can't make energy inside the house vanish without trace, though we can turn it into other forms. So what if one room of the house suddenly starts getting hotter? The heat energy making that happen must be coming from energy that's already inside the house in a different form (maybe it's wood in a fire that's being burned to release the chemical energy locked inside it). If that's not the case, we don't have a closed system: the "extra" heat must be coming into the house from outside (maybe strong sunlight streaming in through the window).

Examples of the conservation of energy

The conservation of energy (and the idea of a "closed system") sounds a bit abstract, but it becomes an awful lot clearer when we consider some real-life examples.

Driving a car

Fill a car up with gasoline and you have a closed system. All the energy you have at your disposal is locked inside the gas in your tank in chemical form. When the gas flows into your engine, it burns with oxygen in the air. The chemical energy in the gas is converted first into heat energy: the burning fuel makes hot expanding gas, which pushes the pistons in the engine cylinders. In this way, the heat is converted into mechanical energy. The pistons turn the crankshaft, gears, and driveshaft and—eventually—the car's wheels. As the wheels turn, they speed the vehicle along the road, giving it kinetic energy (energy of movement).
If a car were 100 percent efficient, all the chemical energy originally locked inside the gasoline would be converted into kinetic energy. Unfortunately, energy is wasted at each stage of this process. Some is lost to friction when metal parts rub and wear against one another and heat up; some energy is lost as sound (cars can be quite noisy—and sound is energy that has to come from somewhere) Not all the energy the car produces moves you down the road: quite a lot has to push against the air (so it's lost to air resistance or drag), while some will be used to power things like the headlights, air conditioning, and so on. Nevertheless, if you measure the energy you start with (in the gasoline) and calculate how much energy you finish with and lose on the way (everything from useful kinetic energy and useless energy lost to friction, sound, air resistance, and so on), you'll find the energy account always balances: the energy you start with is the energy you finish with.

                           Conservation of energy in action: the energy you put into a car, as fuel, is exactly balanced by the useful work the car does plus various energy losses.
Artwork: Like everything else, cars must obey the law of conservation of energy. They convert the energy in fuel into mechanical energy that moves you down the road, but waste quite a lot of energy in the process. If you put 100 units of energy into a car (in the form of fuel), only 15 units or so move you down the road. The rest is wasted as heat losses in the engine (74 percent); parasitic losses (6%, making electricity, for example, to light the headlamps); and drivetrain losses (5%, sending power to the wheels). The 15 useful units of energy are used to overcome drag (air resistance), friction (in the brakes), and rolling resistance (in the tires). Every bit of energy we put into a car has to go somewhere, so the energy outputs (74% + 6% + 5% + 15%) must always exactly add up to the original energy input (100%). Figures for city driving from Where the energy goes, fueleconomy.gov.
Now this only applies if your car is a "closed system." If you're driving along the straight and the road suddenly starts going downhill, you're going to be able to go much further than you'd be able to go otherwise. Does this violate the conservation of energy? No, because we're no longer dealing with a closed system. Your car is gaining kinetic energy from the gasoline in its tank, but it's also gaining kinetic energy because it's going downhill. This isn't a closed system so the conservation of energy doesn't apply anymore.

Boiling a kettle

                                                  A white plastic electric jug kettle.
Photo: An electric kettle like this converts electrical energy into heat energy. 

That's the reverse of the process that happens in the power plant that supplies your home, where electricity is produced using heat energy released by burning a fuel such as coal, oil, or gas.
Boil water with an electric kettle and you're seeing the conservation of energy at work again. Electrical energy drawn from the power outlet on your wall flows into the heating element in the base of your kettle. As the current flows through the element, the element rapidly heats up, so the electrical energy is converted into heat energy that gets passed to the cold water surrounding it. After a couple of minutes, the water boils and (if the power stays on) starts to turn to steam. How does the conservation of energy apply here? Most of the electrical energy that enters the kettle is converted into heat energy in the water, though some is used to provide latent heat of evaporation (the heat we need to give to liquids to turn them into gases such as steam). If you add up the total electrical energy "lost" by the electricity supply and the total energy gained by the water, you should find they're almost exactly the same. Why aren't they exactly equal? Simply because we don't have a closed system here. Some of the original energy is converted to sound and wasted (kettles can be quite noisy). Kettles also give off some heat to their surroundings—so that's also wasted energy.

Pushing a car uphill

In the everyday world, "work" is something you do to earn money; in physics, work has a different meaning. When you do a useful job with a force (a push or a pull), such as moving a car uphill, we say you're doing work, and that takes energy. If you push a car uphill, it has more potential energy at the top of the hill than it had at the bottom. Have you violated the conservation of energy by creating potential energy out of thin air?
No! To push the car, you have to do work against the force of gravity. Your body has to use energy to do work. Most of the energy your body uses is gained by the car as you push it uphill. The energy your body loses is pretty much equal to the work it does against gravity. And the energy the car gains is the same as the work done. So no energy is created or destroyed here: you're simply converting energy stored as fuel inside your body into potential energy stored by the car (because of its height).

Who discovered the conservation of energy?

How do we know the conservation of energy is true? First, it sounds sensible. If you put a heavy log on a fire it might burn for an hour. If you put a second log, roughly the same size, on the fire, it's reasonable to suppose you'll get twice as much heat or the fire will burn twice as long. By the same token, if five bananas can supply your body with an hour's energy, ten bananas should keep you running for two hours—although you might not enjoy guzzling them all at once! In other words, the energy in (the logs you add to the fire or the bananas you eat) is equal to the energy out (the heat you get by burning logs or the energy you make by eating bananas).
                      James Prescott Joule's experiment to demonstrate the mechanical equivalent of heat and the law of conservation of energy.
Photo: The Mechanical Equivalent of Heat: In James Prescott Joule's famous experiment, a falling weight (1) pulls on a rope that passes over a pulley (2). The rope spins an axle (3) that turns a paddle inside a sealed container of water (4). As the paddle spins, the water heats up. Joule proved that the heat energy gained by the water was exactly the same as the potential energy lost by the weight.
... the quantity of heat produced by the friction of bodies, whether solid or liquid, is always proportional to the quantity of force expended.
James Prescott Joule, The Mechanical Equivalent of Heat, 1845.
Reasonable guesswork doesn't quite cut the mustard in science. Really, we need to be sure that the energy we start with in a closed system is the same as the energy we end up with. So how do we know this? One of the first people to confirm the law of conservation of energy experimentally was English physicist James Prescott Joule (1818–1889), who used an ingenious bit of apparatus to find what he called "the mechanical equivalent of heat." He used a falling weight to drive a large paddle wheel sealed inside a container of water. He calculated the potential energy of the weight (the energy it had because of its height above Earth) and reasoned that, as the weight fell, it transferred pretty much all its energy to kinetic energy in the paddle wheel. As the paddle wheel turned, it stirred the water in the container and warmed it up by a small but significant amount. Now we know how much energy it takes to warm a certain mass of water by a certain number of degrees, so Joule was able to figure out how much energy the water had gained. To his delight, he found out that this figure exactly matched the energy lost by the falling weight. Joule's brilliant work on energy was recognized when the international scientific unit of energy (the joule) was named for him.
Joule built on earlier work by Anglo-American physicist Benjamin Thompson (1753–1814), also known as Count Rumford. While working in a Germany artillery factory, Rumford noted that cannon barrels got hot when they were being drilled out. He swiftly realized that the heat was not a magic property of the metal (as many people supposed) but came from the mechanical, frictional process of drilling: the more you drilled, the hotter the metal got. Rumford's simple calculations produced results that, according to Joule, were "not very widely different from that which I have deduced from my own experiments." That was a sign both men were on the right track.

Why perpetual motion machines never work

                             A perpetual motion machine that endlessly cycles water using a waterwheel and Archimedes screw
Back in the 19th century, charlatan inventors would pop up from time to time showing off miracle machines that seemed to be able to drive themselves forever. Inventions like this are called perpetual motion machines: they seem to be able to move forever without anyone adding any more energy. Often, machines like this were blatant tricks: the mechanisms were powered by a concealed assistant who sat in the shadows turning a hidden handle! Some of the machines sound plausible, but all of them unfortunately fall foul of the conservation of energy.
In one version of perpetual motion, illustrated here, water (1) tips down onto a waterwheel, turning it around (2). The turning wheel drives gears (3) that power an Archimedes screw, which lifts the water back up to the top, theoretically allowing the whole cycle to repeat itself forever. Although you might think energy is being recycled as the water moves around, it's also being lost all the time. The water at the top has potential energy and that can indeed drive a waterwheel as it falls. But some energy will be lost to friction as the wheel turns. More energy will be lost to friction in the gears and the screw. So, between them, the wheel, gears, and screw will not have the same amount of energy as the potential energy the water lost originally. That means the screw cannot lift as as much water back to the bath as fell down from it originally—so the machine will very quickly come to a stop.

What about the conservation of mass?

Nuclear reactions seem to create energy out of nothing breaking up or joining together atoms. Do they violate the conservation of energy? No!
Albert Einstein's famous equation E=mc2 shows that energy and mass are different forms of the same thing. Loosely speaking, you can convert a small amount of mass into a large amount of energy (as in a nuclear power plant, where large atoms split apart and give off energy in the process). Einstein's equation shows us we sometimes need to factor mass into the conservation of energy. In a nuclear reaction, we start off with one set of atoms (a certain amount of energy in the form of mass) and end up with a different set of atoms (a different amount of energy locked in their mass) plus energy that's released as heat. If we factor in the mass of the atoms before and after the reaction, plus the energy released in the process, we find the conservation of energy is satisfied exactly. Since mass is a form of energy, it's clear that we can't destroy mass or create it out of nothing in the same way that we can't create or destroy energy.




                                                                          4

⇉Push vs. pull: data movement for linked data structures⇇


As the performance gap between the CPU and main memory continues to grow, techniques to hide memory latency are essential to deliver a high performance computer system. Prefetching can often overlap memory latency with computation for array-based numeric applications. However, prefetching for pointer-intensive applications still remains a challenging problem. Prefetching linked data structures (LDS) is difficult because the address sequence of LDS traversal does not present the same arithmetic regularity as array-based applications and the data dependence of pointer dereferences can serialize the address generation process.

hardware/software mechanism to reduce memory access latencies for linked data structures. Instead of relying on the past address history to predict future accesses, we identify the load instructions that traverse the LDS, and execute them ahead of the actual computation. To overcome the serial nature of the LDS address generation, we attach a prefetch controller to each level of the memory hierarchy and push, rather than pull, data to the CPU. 


Stack (abstract data type)



In computer science, a stack is an abstract data type that serves as a collection of elements, with two principal operations:
  • push, which adds an element to the collection, and
  • pop, which removes the most recently added element that was not yet removed.
The order in which elements come off a stack gives rise to its alternative name, LIFO (last in, first out). Additionally, a peek operation may give access to the top without modifying the stack. The name "stack" for this type of structure comes from the analogy to a set of physical items stacked on top of each other, which makes it easy to take an item off the top of the stack, while getting to an item deeper in the stack may require taking off multiple other items first.

Considered as a linear data structure, or more abstractly a sequential collection, the push and pop operations occur only at one end of the structure, referred to as the top of the stack. This makes it possible to implement a stack as a singly linked list and a pointer to the top element. A stack may be implemented to have a bounded capacity. If the stack is full and does not contain enough space to accept an entity to be pushed, the stack is then considered to be in an overflow state. The pop operation removes an item from the top of the stack.
A stack is needed to implement depth-first search.

         
                                              
                 Similar to a stack of plates, adding or removing is only possible at the top

                           

          
               Simple representation of a stack runtime with push and pop operations.

 

            Hardware stack

A common use of stacks at the architecture level is as a means of allocating and accessing memory.

Basic architecture of a stack

A typical stack, storing local data and call information for nested procedure calls (not necessarily nested procedures). This stack grows downward from its origin. The stack pointer points to the current topmost datum on the stack. A push operation decrements the pointer and copies the data to the stack; a pop operation copies data from the stack and then increments the pointer. Each procedure called in the program stores procedure return information (in yellow) and local data (in other colors) by pushing them onto the stack. This type of stack implementation is extremely common, but it is vulnerable to buffer overflow attacks (see the text).
A typical stack is an area of computer memory with a fixed origin and a variable size. Initially the size of the stack is zero. A stack pointer, usually in the form of a hardware register, points to the most recently referenced location on the stack; when the stack has a size of zero, the stack pointer points to the origin of the stack.
The two operations applicable to all stacks are:
  • a push operation, in which a data item is placed at the location pointed to by the stack pointer, and the address in the stack pointer is adjusted by the size of the data item;
  • a pop or pull operation: a data item at the current location pointed to by the stack pointer is removed, and the stack pointer is adjusted by the size of the data item.
There are many variations on the basic principle of stack operations. Every stack has a fixed location in memory at which it begins. As data items are added to the stack, the stack pointer is displaced to indicate the current extent of the stack, which expands away from the origin.
Stack pointers may point to the origin of a stack or to a limited range of addresses either above or below the origin (depending on the direction in which the stack grows); however, the stack pointer cannot cross the origin of the stack. In other words, if the origin of the stack is at address 1000 and the stack grows downwards (towards addresses 999, 998, and so on), the stack pointer must never be incremented beyond 1000 (to 1001, 1002, etc.). If a pop operation on the stack causes the stack pointer to move past the origin of the stack, a stack underflow occurs. If a push operation causes the stack pointer to increment or decrement beyond the maximum extent of the stack, a stack overflow occurs.
Some environments that rely heavily on stacks may provide additional operations, for example:
  • Duplicate: the top item is popped, and then pushed again (twice), so that an additional copy of the former top item is now on top, with the original below it.
  • Peek: the topmost item is inspected (or returned), but the stack pointer and stack size does not change (meaning the item remains on the stack). This is also called top operation in many articles.
  • Swap or exchange: the two topmost items on the stack exchange places.
  • Rotate (or Roll): the n topmost items are moved on the stack in a rotating fashion. For example, if n=3, items 1, 2, and 3 on the stack are moved to positions 2, 3, and 1 on the stack, respectively. Many variants of this operation are possible, with the most common being called left rotate and right rotate.
Stacks are often visualized growing from the bottom up (like real-world stacks). They may also be visualized growing from left to right, so that "topmost" becomes "rightmost", or even growing from top to bottom. The important feature is that the bottom of the stack is in a fixed position. The illustration in this section is an example of a top-to-bottom growth visualization: the top (28) is the stack "bottom", since the stack "top" (9) is where items are pushed or popped from.
A right rotate will move the first element to the third position, the second to the first and the third to the second. Here are two equivalent visualizations of this process:
apple                         banana
banana    ===right rotate==>  cucumber
cucumber                      apple
cucumber                      apple
banana    ===left rotate==>   cucumber
apple                         banana
A stack is usually represented in computers by a block of memory cells, with the "bottom" at a fixed location, and the stack pointer holding the address of the current "top" cell in the stack. The top and bottom terminology are used irrespective of whether the stack actually grows towards lower memory addresses or towards higher memory addresses.
Pushing an item on to the stack adjusts the stack pointer by the size of the item (either decrementing or incrementing, depending on the direction in which the stack grows in memory), pointing it to the next cell, and copies the new top item to the stack area. Depending again on the exact implementation, at the end of a push operation, the stack pointer may point to the next unused location in the stack, or it may point to the topmost item in the stack. If the stack points to the current topmost item, the stack pointer will be updated before a new item is pushed onto the stack; if it points to the next available location in the stack, it will be updated after the new item is pushed onto the stack.
Popping the stack is simply the inverse of pushing. The topmost item in the stack is removed and the stack pointer is updated, in the opposite order of that used in the push operation.

Stack in main memory

Many CISC-type CPU designs, including the x86, Z80 and 6502, have a dedicated register for use as the call stack stack pointer with dedicated call, return, push, and pop instructions that implicitly update the dedicated register, thus increasing code density. Some CISC processors, like the PDP-11 and the 68000, also have special addressing modes for implementation of stacks, typically with a semi-dedicated stack pointer as well (such as A7 in the 68000). In contrast, most RISC CPU designs do not have dedicated stack instructions and therefore most if not all registers may be used as stack pointers as needed.

Stack in registers or dedicated memory

The x87 floating point architecture is an example of a set of registers organised as a stack where direct access to individual registers (relative the current top) is also possible. As with stack-based machines in general, having the top-of-stack as an implicit argument allows for a small machine code footprint with a good usage of bus bandwidth and code caches, but it also prevents some types of optimizations possible on processors permitting random access to the register file for all (two or three) operands. A stack structure also makes superscalar implementations with register renaming (for speculative execution) somewhat more complex to implement, although it is still feasible, as exemplified by modern x87 implementations.
Sun SPARC, AMD Am29000, and Intel i960 are all examples of architectures using register windows within a register-stack as another strategy to avoid the use of slow main memory for function arguments and return values.
There are also a number of small microprocessors that implements a stack directly in hardware and some microcontrollers have a fixed-depth stack that is not directly accessible. Examples are the PIC microcontrollers, the Computer Cowboys MuP21, the Harris RTX line, and the Novix NC4016. Many stack-based microprocessors were used to implement the programming language Forth at the microcode level. Stacks were also used as a basis of a number of mainframes and mini computers. Such machines were called stack machines, the most famous being the Burroughs B5000.

Applications of stacks

Expression evaluation and syntax parsing

Calculators employing reverse Polish notation use a stack structure to hold values. Expressions can be represented in prefix, postfix or infix notations and conversion from one form to another may be accomplished using a stack. Many compilers use a stack for parsing the syntax of expressions, program blocks etc. before translating into low level code. Most programming languages are context-free languages, allowing them to be parsed with stack based machines.

Backtracking

Another important application of stacks is backtracking. Consider a simple example of finding the correct path in a maze. There are a series of points, from the starting point to the destination. We start from one point. To reach the final destination, there are several paths. Suppose we choose a random path. After following a certain path, we realise that the path we have chosen is wrong. So we need to find a way by which we can return to the beginning of that path. This can be done with the use of stacks. With the help of stacks, we remember the point where we have reached. This is done by pushing that point into the stack. In case we end up on the wrong path, we can pop the last point from the stack and thus return to the last point and continue our quest to find the right path. This is called backtracking.
The prototypical example of a backtracking algorithm is depth-first search, which finds all vertices of a graph that can be reached from a specified starting vertex. Other applications of backtracking involve searching through spaces that represent potential solutions to an optimization problem. Branch and bound is a technique for performing such backtracking searches without exhaustively searching all of the potential solutions in such a space.

Compile time memory management

A number of programming languages are stack-oriented, meaning they define most basic operations (adding two numbers, printing a character) as taking their arguments from the stack, and placing any return values back on the stack. For example, PostScript has a return stack and an operand stack, and also has a graphics state stack and a dictionary stack. Many virtual machines are also stack-oriented, including the p-code machine and the Java Virtual Machine.
Almost all calling conventions‍—‌the ways in which subroutines receive their parameters and return results‍—‌use a special stack (the "call stack") to hold information about procedure/function calling and nesting in order to switch to the context of the called function and restore to the caller function when the calling finishes. The functions follow a runtime protocol between caller and callee to save arguments and return value on the stack. Stacks are an important way of supporting nested or recursive function calls. This type of stack is used implicitly by the compiler to support CALL and RETURN statements (or their equivalents) and is not manipulated directly by the programmer.
Some programming languages use the stack to store data that is local to a procedure. Space for local data items is allocated from the stack when the procedure is entered, and is deallocated when the procedure exits. The C programming language is typically implemented in this way. Using the same stack for both data and procedure calls has important security implications (see below) of which a programmer must be aware in order to avoid introducing serious security bugs into a program.

Efficient algorithms

Several algorithms use a stack (separate from the usual function call stack of most programming languages) as the principle data structure with which they organize their information. These include:

  • Graham scan, an algorithm for the convex hull of a two-dimensional system of points. A convex hull of a subset of the input is maintained in a stack, which is used to find and remove concavities in the boundary when a new point is added to the hull.
  • Part of the SMAWK algorithm for finding the row minima of a monotone matrix uses stacks in a similar way to Graham scan.
  • All nearest smaller values, the problem of finding, for each number in an array, the closest preceding number that is smaller than it. One algorithm for this problem uses a stack to maintain a collection of candidates for the nearest smaller value. For each position in the array, the stack is popped until a smaller value is found on its top, and then the value in the new position is pushed onto the stack.
  • The nearest-neighbor chain algorithm, a method for agglomerative hierarchical clustering based on maintaining a stack of clusters, each of which is the nearest neighbor of its predecessor on the stack. When this method finds a pair of clusters that are mutual nearest neighbors, they are popped and merged.

Security

Some computing environments use stacks in ways that may make them vulnerable to security breaches and attacks. Programmers working in such environments must take special care to avoid the pitfalls of these implementations.
For example, some programming languages use a common stack to store both data local to a called procedure and the linking information that allows the procedure to return to its caller. This means that the program moves data into and out of the same stack that contains critical return addresses for the procedure calls. If data is moved to the wrong location on the stack, or an oversized data item is moved to a stack location that is not large enough to contain it, return information for procedure calls may be corrupted, causing the program to fail.
Malicious parties may attempt a stack smashing attack that takes advantage of this type of implementation by providing oversized data input to a program that does not check the length of input. Such a program may copy the data in its entirety to a location on the stack, and in so doing it may change the return addresses for procedures that have called it. An attacker can experiment to find a specific type of data that can be provided to such a program such that the return address of the current procedure is reset to point to an area within the stack itself (and within the data provided by the attacker), which in turn contains instructions that carry out unauthorized operations.
This type of attack is a variation on the buffer overflow attack and is an extremely frequent source of security breaches in software, mainly because some of the most popular compilers use a shared stack for both data and procedure calls, and do not verify the length of data items. Frequently programmers do not write code to verify the size of data items, either, and when an oversized or undersized data item is copied to the stack, a security breach may occur.

                                        

                                                                     5

⇉  To operation and change operation in differential mathematics (SET) and integral (RESET)⇇

 

In calculus, Leibniz's rule for differentiation under the integral sign, named after Gottfried Leibniz, states that for an integral of the form
where , the derivative of this integral is expressible as
where the partial derivative indicates that inside the integral, only the variation of f(x, t) with x is considered in taking the derivative. Notice that if and are constants rather than functions of , we have a special case of Leibniz's rule:
Besides, if and , which is a common situation as well (for example, in the proof of Cauchy's repeated integration formula), we have:
Thus under certain conditions, one may interchange the integral and partial differential operators. This important result is particularly useful in the differentiation of integral transforms. An example of such is the moment generating function in probability theory, a variation of the Laplace transform, which can be differentiated to generate the moments of a random variable. Whether Leibniz's integral rule applies is essentially a question about the interchange of limits.

General form: Differentiation under the integral sign

Theorem. Let f(x, t) be a function such that both f(x, t) and its partial derivative fx(x, t) are continuous in t and x in some region of the (x, t)-plane, including a(x) ≤ tb(x), x0xx1. Also suppose that the functions a(x) and b(x) are both continuous and both have continuous derivatives for x0xx1. Then, for x0xx1,
This formula is the general form of the Leibniz integral rule and can be derived using the fundamental theorem of calculus. The (first) fundamental theorem of calculus is just the particular case of the above formula where a(x) = a, a constant, b(x) = x, and f(x, t) = f(t).
If both upper and lower limits are taken as constants, then the formula takes the shape of an operator equation:
where is the partial derivative with respect to and is the integral operator with respect to over a fixed interval. That is, it is related to the symmetry of second derivatives, but involving integrals as well as derivatives. This case is also known as the Leibniz integral rule.
The following three basic theorems on the interchange of limits are essentially equivalent:
  • the interchange of a derivative and an integral (differentiation under the integral sign; i.e., Leibniz integral rule);
  • the change of order of partial derivatives;
  • the change of order of integration (integration under the integral sign; i.e., Fubini's theorem).

Three-dimensional, time-dependent case

Figure 1: A vector field F(r, t) defined throughout space, and a surface Σ bounded by curve ∂Σ moving with velocity v over which the field is integrated.
A Leibniz integral rule for a two dimensional surface moving in three dimensional space is
where:
F(r, t) is a vector field at the spatial position r at time t,
Σ is a surface bounded by the closed curve ∂Σ,
dA is a vector element of the surface Σ,
ds is a vector element of the curve ∂Σ,
v is the velocity of movement of the region Σ,
∇⋅ is the vector divergence,
× is the vector cross product,
The double integrals are surface integrals over the surface Σ, and the line integral is over the bounding curve ∂Σ.

Higher dimensions

The Leibniz integral rule can be extended to multidimensional integrals. In two and three dimensions, this rule is better known from the field of fluid dynamics as the Reynolds transport theorem:
where is a scalar function, D(t) and ∂D(t) denote a time-varying connected region of R3 and its boundary, respectively, is the Eulerian velocity of the boundary (see Lagrangian and Eulerian coordinates) and dΣ = n dS is the unit normal component of the surface element.
The general statement of the Leibniz integral rule requires concepts from differential geometry, specifically differential forms, exterior derivatives, wedge products and interior products. With those tools, the Leibniz integral rule in n dimensions is[2]
where Ω(t) is a time-varying domain of integration, ω is a p-form, is the vector field of the velocity, denotes the interior product with , dxω is the exterior derivative of ω with respect to the space variables only and is the time derivative of ω.
However, all of these identities can be derived from a most general statement about Lie derivatives:
Here, the ambient manifold on which the differential form lives includes both space and time.
is the region of integration (a submanifold) at a given instant (it does not depend on , since its parametrization as a submanifold defines its position in time),
is the Lie derivative,
is the spacetime vector field obtained from adding the unitary vector field in the direction of time to the purely spatial vector field from the previous formulas (i.e, is the spacetime velocity of ),
is a diffeomorphism from the one-parameter group generated by the flow of , and
is the image of under such diffeomorphism.
Something remarkable about this form, is that it can account for the case when changes its shape and size over time, since such deformations are fully determined by .

Measure theory statement

Let be an open subset of , and be a measure space. Suppose satisfies the following conditions:
  1. is a Lebesgue-integrable function of for each .
  2. For almost all , the derivative exists for all .
  3. There is an integrable function such that for all and almost every .
Then by the dominated convergence theorem for all ,

Proofs

Proof of basic form

Let
By the definition of the derivative,
Substitute equation (1) into equation (2). The difference of two integrals equals the integral of the difference, and 1/h is a constant, so
Provided that the limit can be passed through the integral sign, we obtain
We claim that the passage of the limit under the integral sign is valid by the bounded convergence theorem (a corollary of the dominated convergence theorem). For each δ > 0, consider the difference quotient
For t fixed, the mean value theorem implies there exists z in the interval [x, x + δ] such that
Continuity of fx(x, t) and compactness of the domain together imply that fx(x, t) is bounded. The above application of the mean value theorem therefore gives a uniform (independent of δ) bound on . The difference quotients converge pointwise to the partial derivative fx by the assumption that the partial derivative exists.
The above argument shows that for every sequence {δn} → 0, the sequence is uniformly bounded and converges pointwise to fx. The bounded convergence theorem states that if a sequence of functions on a set of finite measure is uniformly bounded and converges pointwise, then passage of the limit under the integral is valid. In particular, the limit and integral may be exchanged for every sequence {δn} → 0. Therefore, the limit as δ → 0 may be passed through the integral sign.
For a simpler proof using Fubini's theorem, see the references.

Variable limits form

For a continuous real valued function g of one real variable, and real valued differentiable functions and of one real variable,
This follows from the chain rule and the First Fundamental Theorem of Calculus. Define
,
and
. (The lower limit just has to be some number in the domain of )
Then, can be written as a composition: . The Chain Rule then implies that
.
By the First Fundamental Theorem of Calculus, . Therefore, substituting this result above, we get the desired equation:
.
Note: This form can be particularly useful if the expression to be differentiated is of the form:
Because does not depend on the limits of integration, it may be move out from under the integral sign, and the above form may be used with the Product rule, i.e.

General form with variable limits

Set
where a and b are functions of α that exhibit increments Δa and Δb, respectively, when α is increased by Δα. Then,
A form of the mean value theorem, , where a < ξ < b, may be applied to the first and last integrals of the formula for Δφ above, resulting in
Divide by Δα and let Δα → 0. Notice ξ1a and ξ2b. We may pass the limit through the integral sign:
again by the bounded convergence theorem. This yields the general form of the Leibniz integral rule,

Alternative Proof of General Form with Variable Limits, using the Chain Rule

The general form of Leibniz's Integral Rule with variable limits can be derived as a consequence of the basic form of Leibniz's Integral Rule, the Multivariable Chain Rule, and the First Fundamental Theorem of Calculus. Suppose is defined in a rectangle in the plane, for and . Also, assume and the partial derivative are both continuous functions on this rectangle. Suppose are differentiable real valued functions defined on , with values in (i.e. for every ). Now, set
,   for and
and
,   for
Then, by properties of Definite Integrals, we can write
Since the functions are all differentiable (see the remark at the end of the proof), by the Multivariable Chain Rule, it follows that is differentiable, and its derivative is given by the formula:
 
Now, note that for every , and for every , we have that , because when taking the partial derivative with respect to of , we are keeping fixed in the expression ; thus the basic form of Leibniz's Integral Rule with constant limits of integration applies. Next, by the First Fundamental Theorem of Calculus, we have that ; because when taking the partial derivative with respect to of , the first variable is fixed, so the fundamental theorem can indeed be applied.
Substituting these results into the equation for above gives:
as desired.
There is a technical point in the proof above which is worth noting: applying the Chain Rule to requires that already be Differentiable. This is where we use our assumptions about . As mentioned above, the partial derivatives of are given by the formulas and . Since is continuous, its integral is also a continuous function, and since is also continuous, these two results show that both the partial derivatives of are continuous. Since continuity of partial derivatives implies differentiability of the function, is indeed differentiable.

Three-dimensional, time-dependent form

At time t the surface Σ in Figure 1 contains a set of points arranged about a centroid . The function can be written as
with independent of time. Variables are shifted to a new frame of reference attached to the moving surface, with origin at . For a rigidly translating surface, the limits of integration are then independent of time, so:
where the limits of integration confining the integral to the region Σ no longer are time dependent so differentiation passes through the integration to act on the integrand only:
with the velocity of motion of the surface defined by
This equation expresses the material derivative of the field, that is, the derivative with respect to a coordinate system attached to the moving surface. Having found the derivative, variables can be switched back to the original frame of reference. We notice that (see article on curl)
and that Stokes theorem equates the surface integral of the curl over Σ with a line integral over ∂Σ:
The sign of the line integral is based on the right-hand rule for the choice of direction of line element ds. To establish this sign, for example, suppose the field F points in the positive z-direction, and the surface Σ is a portion of the xy-plane with perimeter ∂Σ. We adopt the normal to Σ to be in the positive z-direction. Positive traversal of ∂Σ is then counterclockwise (right-hand rule with thumb along z-axis). Then the integral on the left-hand side determines a positive flux of F through Σ. Suppose Σ translates in the positive x-direction at velocity v. An element of the boundary of Σ parallel to the y-axis, say ds, sweeps out an area vt × ds in time t. If we integrate around the boundary ∂Σ in a counterclockwise sense, vt × ds points in the negative z-direction on the left side of ∂Σ (where ds points downward), and in the positive z-direction on the right side of ∂Σ (where ds points upward), which makes sense because Σ is moving to the right, adding area on the right and losing it on the left. On that basis, the flux of F is increasing on the right of ∂Σ and decreasing on the left. However, the dot product v × F • ds = −F × vds = −F • v × ds. Consequently, the sign of the line integral is taken as negative.
If v is a constant,
which is the quoted result. This proof does not consider the possibility of the surface deforming as it moves.

Alternative derivation

Lemma. One has:
Proof. From proof of the fundamental theorem of calculus,
and
Suppose a and b are constant, and that f(x) involves a parameter α which is constant in the integration but may vary to form different integrals. Assume that f(x, α) is a continuous function of x and α in the compact set {(x, α) : α0 ≤ α ≤ α1 and axb}, and that the partial derivative fα(x, α) exists and is continuous. If one defines:
then may be differentiated with respect to α by differentiating under the integral sign, i.e.,
By the Heine–Cantor theorem it is uniformly continuous in that set. In other words, for any ε > 0 there exists Δα such that for all values of x in [a, b],
On the other hand,
Hence φ(α) is a continuous function.
Similarly if exists and is continuous, then for all ε > 0 there exists Δα such that:
Therefore,
where
Now, ε → 0 as Δα → 0, so
This is the formula we set out to prove.
Now, suppose
where a and b are functions of α which take increments Δa and Δb, respectively, when α is increased by Δα. Then,
A form of the mean value theorem, where a < ξ < b, can be applied to the first and last integrals of the formula for Δφ above, resulting in
Dividing by Δα, letting Δα → 0, noticing ξ1a and ξ2b and using the above derivation for
yields
This is the general form of the Leibniz integral rule.

Examples

General examples

Example 1

Consider the function
The function under the integral sign is not continuous at the point (x, α) = (0, 0), and the function φ(α) has a discontinuity at α = 0 because φ(α) approaches ±Ï€/2 as α → 0±.
If we differentiate φ(α) with respect to α under the integral sign, we get
which is, of course, true for all values of α except α = 0. This may be integrated (with respect to α) to find

Example 2

An example with variable limits:

Examples for evaluating a definite integral

Example 3

The principle of differentiating under the integral sign may sometimes be used to evaluate a definite integral. Consider:
Now,
As x varies from 0 to π, we have
Hence,
Therefore,
Integrating both sides with respect to α, we get:
C1 = 0 follows from evaluating φ(0):
To determine C2 in the same manner, we should need to substitute in a value of α greater than 1 in φ(α). This is somewhat inconvenient. Instead, we substitute α = 1/β, where |β| < 1. Then,
Therefore, C2 = 0.
The definition of φ(α) is now complete:
The foregoing discussion, of course, does not apply when α = ±1, since the conditions for differentiability are not met.

Example 4

First we calculate:
The limits of integration being independent of a, we have:
On the other hand:
Equating these two relations then yields
In a similar fashion, pursuing yields
Adding the two results then produces
which computes as desired.
This derivation may be generalized. Note that if we define
it can easily be shown that
Given I1, this integral reduction formula can be used to compute all of the values of In for n > 1.

Example 5

Here, we consider the integral
Differentiating under the integral with respect to α, we have
Therefore:
However, by definition, I(π/2) = 0, hence C = π2/8 and

Example 6

Here, we consider the integral
We introduce a new variable φ and rewrite the integral as
When φ = 1 this equals the original integral. However, this more general integral may be differentiated with respect to φ:
This is the line integral of over the unit circle. By Green's Theorem, it equals the double integral over the unit disk of which equals 0. This implies that f(φ) is constant. The constant may be determined by evaluating f at φ = 0:
Therefore, the original integral also equals 2Ï€.

Other problems to solve

There are innumerable other integrals that can be solved using the technique of differentiation under the integral sign. For example, in each of the following cases, the original integral may be replaced by a similar integral having a new parameter α:
The first integral, the Dirichlet integral, is absolutely convergent for positive α but only conditionally convergent when α is 0. Therefore, differentiation under the integral sign is easy to justify when α > 0, but proving that the resulting formula remains valid when α is 0 requires some careful work.

Applications to series

The measure-theoretic version of differentiation under the integral sign also applies to summation (finite or infinite) by interpreting summation as counting measure. An example of an application is the fact that power series are differentiable in their radius of convergence.

In popular culture

Differentiation under the integral sign is mentioned in the late physicist Richard Feynman's best-selling memoir Surely You're Joking, Mr. Feynman! in the chapter "A Different Box of Tools". He describes learning it, while in high school, from an old text, Advanced Calculus (1926), by Frederick S. Woods (who was a professor of mathematics in the Massachusetts Institute of Technology). The technique was not often taught when Feynman later received his formal education in calculus, but using this technique, Feynman was able to solve otherwise difficult integration problems upon his arrival at graduate school at Princeton University:
One thing I never did learn was contour integration. I had learned to do integrals by various methods shown in a book that my high school physics teacher Mr. Bader had given me. One day he told me to stay after class. "Feynman," he said, "you talk too much and you make too much noise. I know why. You're bored. So I'm going to give you a book. You go up there in the back, in the corner, and study this book, and when you know everything that's in this book, you can talk again." So every physics class, I paid no attention to what was going on with Pascal's Law, or whatever they were doing. I was up in the back with this book: "Advanced Calculus", by Woods. Bader knew I had studied "Calculus for the Practical Man" a little bit, so he gave me the real works—it was for a junior or senior course in college. It had Fourier series, Bessel functions, determinants, elliptic functions—all kinds of wonderful stuff that I didn't know anything about. That book also showed how to differentiate parameters under the integral sign—it's a certain operation. It turns out that's not taught very much in the universities; they don't emphasize it. But I caught on how to use that method, and I used that one damn tool again and again. So because I was self-taught using that book, I had peculiar methods of doing integrals. The result was, when guys at MIT or Princeton had trouble doing a certain integral, it was because they couldn't do it with the standard methods they had learned in school. If it was contour integration, they would have found it; if it was a simple series expansion, they would have found it. Then I come along and try differentiating under the integral sign, and often it worked. .


                                                                      6

                                  ⇉   SPACE AND TIME on PUSH AND PULL ⇇

 

Both push and pull drive our galaxy's race through space


Discovery of the 'dipole repeller' confirms that both attraction and repulsion are at play in our extragalactic neighborhood   

Although we can't feel it, we're in constant motion: the earth spins on its axis at about 1,600 km/h; it orbits around the sun at about 100,000 km/h; the sun orbits our Milky Way galaxy at about 850,000 km/h; and the Milky Way galaxy and its companion galaxy Andromeda are moving with respect to the expanding universe at roughly 2 million km/h (630 km per second). But what is propelling the Milky Way's race through space?

Until now, scientists assumed that a dense region of the universe is pulling us toward it, in the same way that gravity made Newton's apple fall to earth. The initial "prime suspect" was called the Great Attractor, a region of a half dozen rich clusters of galaxies 150 million lightyears from the Milky Way. Soon after, attention was drawn to an area of more than two dozen rich clusters, called the Shapley Concentration, which sits 600 million lightyears beyond the Great Attractor.

By identifying the Dipole Repeller, the researchers were able to reconcile both the direction of the Milky Way's motion and its magnitude. They expect that future ultra-sensitive surveys at optical, near-infrared and radio wavelengths will directly identify the few galaxies expected to lie in this void, and directly confirm the void associated with the Dipole Repeller.   

Earth is a time space dimension, so it operates in a way that what fills the space then creates either distance or nearness, a longer time or a shorter time.

Congruence is an amazing word, when you allow the vibration of it to truly sink in. It describes a situation where in the moment (time), you are filling your cells and being with a message that all parts of you is agreeing on (space). If you understand this, then understand that everything that occurs as a response to you, offers you a mirror. For example, you’re a full yes to something, and then you present the idea to someone else and they’re a full no. Do you still remain a full yes? Or do you see it for what it is, you are in congrunece in this moment, and they are not? Do you use this as an opportunity to sharpen your articulation, so they are able to get a clear idea of just how congruent you are? Or do you slink away wounded, contracted and caved in? In response to a no do you strengthen your yes, so that you’re not just in agreement you are also amplified? 

If even a small part of you is out of alignment with the rest of you, then over time it will make itself known. That small part will convince you that it’s there to look after your best interest and keep you safe. You can consider it to be your gate keeper of sorts, it wants you to put your money where your mouth is. It will niggle, and niggle, and niggle, and niggle, in a quest for you to make the same decision over, and over, and over, and over. Eventually when enough of you is amplified you’ll  have pulled the keeper to your side, the gate opens and this becomes your new way of being. Instantly you feel the change because the world begins to respond very differently to you. That part of your life gets easier, just for a little while.

So pay attention to the areas where you’re broadcasting and amplifying a push pull type of energy. Sending mixed messages into the universe, also making it such that you’re not amplifed enough to become magnetic. You stengthen, then weaken, then weaken, then weaken, then strengthen. The energy is counter productive. Empower your stance, empower your decision and empower your actions.

 

  +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

                                Gen . Mac Tech in operation to change operation 

       
                          Hasil gambar untuk aircraft galaxy space

+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++