Thursday, July 26, 2018

Parts of a roof truss



A roof truss is an engineered panel made up of triangular parts. Set out below are the main structural components and fixing points in a standard 'A' type truss.
  1. Apex
  2. Apex plate
  3. Top chord
  4. Heel plate
  5. 1/3 point plate
  6. Bottom chord
  7. Slice plate
  8. Heel
  9. 1/4 point plate
  10. Web
  11. Nominal span
  12. Overhang
Parts of a roof truss.

Top chord

The top chord performs the job of a rafter in a conventional roof. It carries the tile battens or sheet roofing battens, and is set at the pitch or angle of the roof. In the example above, there are two top chords which meet at the apex.

Bottom chord

The bottom chord is connected at each end to the top chords using heel plates. It main structural function is to stop the top chords from spreading apart when they are under a load. It also serves as aceiling joist, allowing the ceiling lining to be fixed to the underside.

Webs

The webs are the internal members that run between the top and bottom chords. They help to give the truss its strength and rigidity by transferring the stresses in the chords throughout the structure. This is what enables a truss to span the full width of a building using small cross-sectional members.

Joints

The chords and webs are joined at the panel points by nail plates. The plates are galvanised steel sheets that have spikes protruding on one side. When they are pressed into the timber, with one plate on each side of the join, they form a solid fixing that is very strong when the truss is in an upright position.

Negative Effects of Computers and the Internet on Society

Computers and the Internet have touched almost all aspects of life. It is rare to come across a business or household that does not experience routine use of a computer in some shape or form.
Technology has allowed people to have higher levels of convenience and proficiency. Many people today would find it very difficult to go back to an age where computers were not in existence.
In addition, society has become accustomed to on-demand answers or solutions to requests or services and the Internet is the platform which fulfills this need. These are some of the positive effects of technology on society.
While there have been many positive effects of computers on society, there have also been some drawbacks too. Issues such as security and complacency have increased in addition to society's ever growing dependence on computers.

Let's take a look at some of the positive and negative effects of computers and the Internet on society:Computers and the Internet have touched almost all aspects of life. It is rare to come across a business or household that does not experience routine use of a computer in some shape or form.
Technology has allowed people to have higher levels of convenience and proficiency. Many people today would find it very difficult to go back to an age where computers were not in existence.
In addition, society has become accustomed to on-demand answers or solutions to requests or services and the Internet is the platform which fulfills this need. These are some of the positive effects of technology on society.
While there have been many positive effects of computers on society, there have also been some drawbacks too. Issues such as security and complacency have increased in addition to society's ever growing dependence on computers.
Let's take a look at some of the positive and negative effects of computers and the Internet on society:
Negative Effects
Unfortunately despite all the positives associated with computers and the Internet, there are some drawbacks too. These are issues society has to contend with in order to achieve the benefits and often trade-offs have to be made.
Security is one of the most prominent negative effects which emerges with the use of technology. The criminal element in society has found many ways to exploit and harm others by using computers and the Internet as a weapon instead of the tool it was designed to be.
Crimes such as identity theft, hacking, embezzlement, and other kinds of monetary theft have increased the risks of doing business online, and these have to be mitigated through using software and being vigilant. These concerns should not deter people from using the Internet, but it is a real concern which must be dealt with.
Complacency is another negative effect. While computers and the Internet have enhanced quality of life, sometimes the question begs asking of whether or not society has become too dependent on computers instead of thinking for one's self. Many people operate on the assumption the computer is always right, and this can be a dangerous notion.
While computers themselves don't make mistakes, the human design behind the software can and do make mistakes, nothing is 100% infallible.
Programmers, while in most cases are pretty accurate, do have typos or software can contain glitches. Since technology is essentially tied to everything from banking, parking meters, health insurance, and medical care, it is important to be vigilant and if something seems off to always question it.
This complacency leads to dependence. Are computers doing too much "thinking" for people? Today many people have no idea of how to manually do transactions or activities that computers routinely take care of these days.

AVERAGE TOTAL COST CURVE:

A curve that graphically represents the relation between average total cost incurred by a firm in the short-run product of a good or service and the quantity produced. The average total cost curve is constructed to capture the relation between average total cost and the level of output, holding other variables, like technology and resource prices, constant. The average total cost curve is one of three average curves. The other two are average variable cost curve and average fixed cost curve. A related curve is the marginal cost curve.
The average total cost curve is U-shaped. Average total cost is relatively high for small quantities of output, then as production increases, it declines, reaches a minimum value, then rises.

Because average total cost is a combination of average variable cost andaverage fixed cost, the U-shape of the average total cost curve is a result of both underlying averages. At small production quantities, both average fixed cost and average variable cost decline, resulting in a negatively-sloped average total cost curve.
However, because of the law of diminishing marginal returns, average variable cost eventually increases, which overwhelms the continuing decline of average fixed cost and results in a positively-sloped average total cost curve.

Average Total Cost Curve


The graph to the right is the average total cost curve for theshort-run production of Wacky Willy Stuffed Amigos (those cute and cuddly armadillos and tarantulas). The quantity of Stuffed Amigos production, measured on the horizontal axis, ranges from 0 to 10 and the average total cost incurred in the production of Stuffed Amigos, measured on the vertical axis, ranges from about $3 to over $8.
As noted above, the average total cost curve is U-shaped. For the first 6 Stuffed Amigos, average total cost declines from over $8 to a low of $3. However, for the production beyond 7 Stuffed Amigos, average total cost increases.
While it would be easy to attribute the U-shape of the average total cost curve to increasing, then decreasing marginal returns (and the law of diminishing marginal returns), such is not completely true. While the law of diminishing marginal returns is indirectly responsible for the positively-sloped portion of the average total cost curve, the negatively-sloped portion is attributable to increasing marginal returns, and perhaps more importantly to declining average fixed cost.


The average total cost curve is most important to the analysis of a firm's short-run production when compared to the price. If price is greater than average total cost, then a firm receives economic profit on each unit of the output produced and sold. If price is less than average total cost, then a firm incurs a loss on each unit produced and sold. However, whether or not the loss is great enough to force the firm to shut down production depends on a comparison between price and average variable cost.

potential benefits from monopoly

The debate about monopoly will never be settled! 

The consensus seems to be that the economic case for and against monopoly needs to be judged on a case by case basis  - particularly when assessing the impact on economic welfare.
The standard economic case against monopoly is that, with the same cost structure, a monopoly supplier will produce at a lower output and charge a higher price than a competitive industry. This leads to a net loss of economic welfare and efficiency because price is driven above marginal cost - leading to allocative inefficiency.
The diagram below shows how price and output differ between a competitive and a monopolistic industry. We have assumed that the cost structure for both the competitive firm and the monopoly is the same - indeed we have assumed that output can be supplied at a constant marginal and average cost.
Assuming that the monopolist seeks to maximise profits and that they take the whole of the market demand curve, then the price under monopoly will be higher and the output lower than the competitive market equilibrium.
This leads to a deadweight loss of consumer surplus and therefore a loss of static economic efficiency.

CAN MONOPOLY BE DEFENDED?
Monopoly and Economies of Scale
Because monopoly producers are often supplying goods and services on a very large scale, they may be better placed to take advantage of economies of scale - leading to a fall in the average total costs of production. These reductions in costs will lead to an increase in monopoly profits but some of the gains in productive efficiency might be passed onto consumers in the form of lower prices. The effect of economies of scale is shown in the diagram above.
Economies of scale provide potential gains in economic welfare for both producers and consumers.


Regulation of monopoly
Because of the potential economic welfare loss arising from the exploitation of monopoly power, the Government regulates some monopolies. Regulators can control annual price increases and introduce fresh competition into particular industries


Monopoly and Innovation

 (Research and Development)

How are the supernormal profits of monopoly used?

 

Is consumer surplus of equal value to producer surplus?

Are large-scale firms required to create a comparative advantage in global markets? Some economists argue that large-scale firms are required to be competitive in international markets.

 

 

However some of the supernormal profits might be used to invest in research and development programmes that have the potential to bring dynamic efficiency gains to consumers in the markets. There is a continuing debate about whether competitive or monopolistic markets provide the best environment for high levels of research spending.


Price Discrimination

Are there potential welfare improvements from price discrimination? Some forms of price discrimination benefit certain consumers.



Domestic monopoly but international competition
A firm may have substantial domestic monopoly power but face intensive competition from overseas producers. This limits their market power and helps keep prices down for consumers. A good example to use here would be the domestic steel industry. Corus produces most of the steel manufactured inside the UK but faces intensive competition from overseas steel producers.


Contestable markets!
Contestable market theory predicts that monopolists may still be competitive even if they enjoy a dominant position in their market. Their price and output decisions will be affected by the threat of "hit and run entry" from other firms if they allow their costs to rise and inefficiencies to develop. 

Organic Compounds

The chemical compounds of living things are known as organic compounds because of their association with organisms. Organic compounds, which are the compounds associated with life processes, are the subject matter of organic chemistry. Among the numerous types of organic compounds, four major categories are found in all living things: carbohydrates, lipids, protein, and nucleic acids.

Carbohydrates

Almost all organisms use carbohydrates as sources of energy. In addition, some carbohydrates serve as structural materials. Carbohydrates are molecules composed of carbon, hydrogen, and oxygen; the ratio of hydrogen atoms to oxygen atoms is 2:1.


Simple carbohydrates, commonly referred to as sugars, can be monosaccharides if they are composed of single molecules, or disaccharides if they are composed of two molecules. The most important monosaccharide is glucose, a carbohydrate with the molecular formula C6H12O6. Glucose is the basic form of fuel in living things. It is soluble and is transported by body fluids to all cells, where it is metabolized to release its energy. Glucose is the starting material for cellular respiration, and it is the main product of photosynthesis.
1 shows that in the synthesis of sucrose, a water molecule is produced. The process is therefore called a dehydration. The reversal of the process is hydrolysis, a process in which the molecule is split and the elements of water are added.) Lactose is composed of glucose and galactose units. 




Figure 1
Glucose and fructose molecules combine to form the disaccharide sucrose.
Complex carbohydrates are known as polysaccharides. Polysaccharides are formed by linking innumerable monosaccharides. Among the most important polysaccharides are the starches, which are composed of hundreds or thousands of glucose units linked to one another. Starches serve as a storage form for carbohydrates. Much of the world's human population satisfies its energy needs with the starches of rice, wheat, corn, and potatoes.
Two other important polysaccharides are glycogen and cellulose. Glycogen is also composed of thousands of glucose units, but the units are bonded in a different pattern than in starches. Glycogen is the form in which glucose is stored in the human liver. Cellulose is used primarily as a structural carbohydrate. It is also composed of glucose units, but the units cannot be released from one another except by a few species of organisms. Wood is composed chiefly of cellulose, as are plant cell walls. Cotton fabric and paper are commercial cellulose products.

Lipids

Lipids are organic molecules composed of carbon, hydrogen, and oxygen atoms. The ratio of hydrogen atoms to oxygen atoms is much higher in lipids than in carbohydrates. Lipids include steroids (the material of which many hormones are composed), waxes, and fats.
2 ). A glycerol molecule contains three hydroxyl (—OH) groups. A fatty acid is a long chain of carbon atoms (from 4 to 24) with a carboxyl (—COOH) group at one end. The fatty acids in a fat may be all alike or they may all be different. They are bound to the glycerol molecule by a process that involves the removal of water. 


Arithmetic Operations and Functions

Operations

In FORTRAN, addition and subtraction are denoted by the usual plus (+) and minus (-) signs. Multiplication is denoted by an asterisk (*). This symbol must be used to denote every multiplication; thus to multiply N by 2, we must use 2 * N or N * 2not 2N. Division is denoted by a slash (/), and exponentiation is denoted by a pair of asterisks (**).

OperatorOperation
+addition, unary plus
-subtraction, unary minus
*multiplication
/division
**exponentiation

Real Arithmetic

Providing all variables and constants in the expression are real, real arithmetic will be carried out as expected, with no decimal places being truncated.

Integer Arithmetic

Providing the expression has all integers, subtraction, addition, multiplication and exponentiation will prove no problem. However integer division is somewhat different than normal division with real values. Integer division ignores the fractional part. Any decimal places are truncated.

Example

5 / 2 gives the result 2 instead of 2.5

3 / 4 gives the result 0 instead of 0.75

Mixed Mode Arithmetic

Mixed mode arithmetic is when an expression contains both reals and integers. If ANY of the operands are real then result of the operation will be real. However, mixed mode arithmetic should be used with extreme care. You may think you have a real operand when in reality you have two integer operands.

Example

5 / 2 * 3.0is 6.0 Incorrect because the order of operation is left to right. 5/2 = 2 then 2 * 3.0 = 6.0
3.0 * 5 / 2is 7.5 Correct because of mixed mode arithmetic 3.0 * 5 = 15.0 then 15.0/2 = 7.5

Mixed Mode Variable Assignments

If the variable to which an expression is assigned has been declared as a real variable, any decimal places resulting from the evaluation of the expression will be preserved.

Example

real variable 5 * 2.1 will have a value of 10.5.

However, if the variable to which an expression is assigned has been declared as an integer variable, any decimal places resulting from the evaluation of the expression will be lost.

Example

integer variable 5 * 2.1 will have a value of 10

Priority Rules.

Arithmetic expressions are evaluated in accordance with the following priority rules:

  • All exponentiations are performed first; consecutive exponentiations are performed from right to left.
  • All multiplication and divisions are performed next, in the order in which they appear from left to right.
  • The additions and subtractions are performed last, in the order in which they appear from left to right.

Functions

FORTRAN provides several intrinsic functions to carry out calculations on am number of values or arguments and return as result. Commonly used functions are shown in the table below. To use a function we simply give the function name followed by the argument(s) enclosed in parenthesis.

   
   funtionname (name1, name2,.......)
   

Some FORTRAN Functions





FunctionDescriptionType of Argument(s)*Type of Value
ABS (x)Absolute value of xI, R, DPSame as argument
COS (x)Cosine of x radiansR, DPSame as argument
DBLE(x)Conversion of x to double precision formI, RDP
DPROD(x,y)Double precision product of x and yRDP
EXP(x)Exponential functionR, DPSame as argument
INT(x)Integer part of xR, DPI
LOG(x)Natural logarithm of xR, DPSame as argument
MAX(xl, . . . , Xn)Maximum of xl, . . .,xnI, R, DPSame as argument
MIN(xl, . . . , xn)Minimum of xl, . . ., xnI, R, DPSame as argument
MOD(x,y)x (mod y); x - INT(x/y) * yI, R, DPSame as argument
NINT(x)x rounded to nearest integerR, DPI
REAL(x)Conversion of x to real typeI, DPR
SIN(x)Sine of x radiansR, DPSame as argument

Angle of repose

The angle of repose or the critical angle of repose,of a granular material is the steepest angle of descent or dip relative to the horizontal plane to which a material can be piled without slumping. At this angle, the material on the slope face is on the verge of sliding. The angle of repose can range from 0° to 90°. Smooth, rounded sand grains cannot be piled as steeply as can rough, interlocking sands. If a small amount of water is able to bridge the gaps between particles, electrostatic attraction of the water to mineral surfaces will increase soil strength.


When bulk granular materials are poured onto a horizontal surface, a conical pile will form. The internal angle between the surface of the pile and the horizontal surface is known as the angle of repose and is related to the density, surface area and shapes of the particles, and the coefficient of friction of the material. However, a 2011 study shows that the angle of repose is also gravity-dependent. Material with a low angle of repose forms flatter piles than material with a high angle of repose.  


Contents

  • 1 Applications of theory
  • 2 Measurement
  • 3 Exploitation by antlion and wormlion (Vermileonidae) larvae
  • 4 Methods in determining the angle of repose
    • 4.1 Tilting box method
    • 4.2 Fixed funnel method
    • 4.3 Revolving cylinder method 
  • 5 Angle of repose of various materials

    Applications of theory 



  • The angle of repose is sometimes used in the design of equipment for the processing of particulate solids. For example, it may be used to design an appropriate hopper or silo to store the material, or to size a conveyor belt for transporting the material. It can also be used in determining whether or not a slope (of a stockpile, or uncompacted gravel bank, for example) will likely collapse; thetalus slope is derived from angle of repose and represents the steepest slope a pile of granular material will take. This angle of repose is also crucial in correctly calculating stability in vessels.
    It is also commonly used by mountaineers as a factor in analysing avalanchedanger in mountainous areas.

    Measurement

    There are numerous methods for measuring angle of repose and each produces slightly different results. Results are also sensitive to the exact methodology of the experimenter. As a result, data from different labs are not always comparable. One method is the triaxial shear test, another is the direct shear test.
    If the coefficient of static friction is known of a material, then a good approximation of the angle of repose can be made with the following function. This function is somewhat accurate for piles where individual objects in the pile are minuscule and piled in random order.
    \tan{(\theta)} \approx \mu_\mathrm{s}\,
    where, μs is the coefficient of static friction, and θ is the angle of repose. 

    Exploitation by antlion and wormlion (Vermileonidae) larvae  



  • The larvae of the antlions and the unrelated wormlions Vermileonidae trap small insects such as ants by digging conical pits in loose sand, such that the slope of the walls is effectively at the critical angle of repose for the sand. They achieve this by flinging the loose sand out of the pit and permitting the sand to settle at its critical angle of repose as it falls back. Thus, when a small insect, commonly an ant, blunders into the pit, its weight causes the sand to collapse below it, drawing the victim toward the center where the predator that dug the pit lies in wait under a thin layer of loose sand. The larva assists this process by vigorously flicking sand out from the center of the pit when it detects a disturbance. This undermines the pit walls and causes them to collapse toward the center. The sand that the larva flings also pelts the prey with so much loose, rolling material as to prevent it from getting any foothold on the easier slopes that the initial collapse of the slope has presented. The combined effect is to bring the prey down to within grasp of the larva, which then can inject venom and digestive fluids.

Nervous tissues

General

All living cells have the ability to react to stimuli. Nervous tissue is specialised to react to stimuli and to conduct impulses to various organs in the body which bring about a response to the stimulusNerve tissue (as in the brain, spinal cord and peripheral nerves that branch throughout the body) are all made up of specialised nerve cells called neurons. Neurons are easily stimulated and transmit impulses very rapidly. A nerve is made up of many nerve cell fibres (neurons) bound together by connective tissue. A sheath of dense connective tissue, the epineurium surrounds the nerve. This sheath penetrates the nerve to form the perineurium which surrounds bundles of nerve fibres. blood vessels of various sizes can be seen in the epineurium. Theendoneurium, which consists of a thin layer of loose connective tissue, surrounds the individual nerve fibres.



Structure of a Motor Neuron

A motor neuron has many processes (cytoplasmic extensions), called dendtrites, which enter a large, grey cell body at one end. A single process, theaxon, leaves at the other end, extending towards the dendrites of the next neuron or to form a motor endplate in a muscle. Dendrites are usually short and divided while the axons are very long and does not branched freely. The impulses are transmitted through the motor neuron in one direction, i.e.into the cell body by the dendrites and away from the cell body by the axon . The cell body is enclosed by a cell (plasma) membrane and has acentral nucleusGranules, called Nissl, bodies are found in the cytoplasm of the cell body. Within the cell body, extremely fine neurofibrils extend from the dendrites into the axon. The axon is surrounded by the myelin sheath, which forms a whitish, non-cellular, fatty layer around the axon. Outside the myelin sheath is a cellular layer called the neurilemma or sheath of Schwann cells. The myelin sheath together with the neurilemma is also known as the medullary sheath. This medullary sheath is interrupted at intervals by the nodes of Ranvier.



A motor neuron
Nerve cells are functionally connected to each other at a junction known as a synapse, where the terminal branches of an axon and the dendrites of another neuron lie in close proximity to each other but never make direct contact.



A Synapse

Classification of Neurons

On the basis of their structure, neurons can also be classified into three main types:

  • Unipolar Neurons.Sensory neurons have only a single process or fibre which divides close to the cell body into two main branches (axon and dendrite). Because of their structure they are often referred to as unipolar neurons.
  • Multipolar Neurons.Motor neurons, which have numerous cell processes (an axon and many dendrites) are often referred to as multipolar neurons. Interneurons are also multipolar.
  • Bipolar Neurons.Bipolar neurons are spindle-shaped, with a dendrite at one end and an axon at the other . An example can be found in the light-sensitive retina of the eye.
A diagram showing the different neurons

Functions of Nerve Tissue

  • Nervous tissue allows an organism to sense stimuli in both the internal and external environment.
  • The stimuli are analysed and integrated to provide appropriate, co-ordinated responses in various organs.
  • The afferent or sensory neurons conduct nerve impulses from the sense organs and receptors to the central nervous system.
  • Internuncial or connector neurons supply the connection between the afferent and efferent neurons as well as different parts of the central nervous system.
  • Efferent or somatic motor neurons transmit the impulse from the central nervous system to a muscle (the effector organ) which then react to the initial stimulus.
  • Autonomic motor or efferent neurons transmit impulses to the involuntary muscles and glands.

Defining High, Mid and Low-Level Languages

I’ve been writing quite a bit recently about the differences between languages.  Mostly I’ve just been whining about how annoying it is that everyone keeps searching for the “one language to rule them all”, the Aryan Language if you will.  Over the course of some of these articles, I’ve made some rather loosely defined references to terms like “general purpose” and “mid-level” when trying to describe these languages. 
Several people have (rightly) called me out on these terms, arguing that I haven’t really defined what they mean, so I shouldn’t be using them to try to argue a certain point.  In the case of “general purpose language”, I have to admit that I tend to horribly misuse the term and any instances within my writing should be discarded without thought.  However, I think with a little bit of reflection, we can come to some reasonable definitions for high-, mid- and low-level languages. 
To that end, I present the “Language Spectrum of Science!” (cue reverb)

Language Spectrum of Science 
This scale is admittedly arbitrary and rather loosely defined in and of itself, but I think it should be a sufficient visual aid in conveying my point.  In case you hadn’t guessed, red languages are low-level, green languages are high-level and that narrow strip of yellow represents the mid-level languages.  Obviously I’m leaving out a large number of languages which could be represented with equal validity, but I only have a finite number of pixels in page-width.It’s also important to note that languages aren’t really points on the spectrum, but rather they span ranges which are more or less wide, depending on the capabilities.  These ranges may overlap considerably (as in the case of Java and Scala) or may be entirely disjoint (Assembly and Ruby).  In short, the scale is somewhat blurry and shouldn’t be taken as a canonical reference.

Low-Level

Of all of the categories, it’s probably easiest to define what it means to be a low-level language.  Machine code is low level because it runs directly on the processor.  Low-level languages are appropriate for writing operating systems or firmware for micro-controllers.  They can do just about anything with a little bit of work, but obviously you wouldn’t want to write the next major web framework in one of them (I can see it now, “Assembly on Rails”).

Characteristics

  • Direct memory management
  • Little-to-no abstraction from the hardware
  • Register access
  • Statements usually have an obvious correspondence with clock cycles
  • Superb performance
C is actually a very interesting language in this category (more so C++) because of how broad its range happens to be.  C allows you direct access to registers and memory locations, but it also has a number of constructs which allow significant abstraction from the hardware itself.  Really, C and C++ probably represent the most broad spectrum languages in existence, which makes them quite interesting from a theoretical standpoint.  In practice, both C and C++ are too low-level to do anything “enterprisy”.

Mid-Level

This is where things start getting vague.  Most high-level languages are well defined, as are low-level languages, but mid-level languages tend to be a bit difficult to box.  I really define the category by the size of application I would be willing to write using a given language.  I would have no problem writing and maintaining a large desktop application in a mid-level language (such as Java), whereas to do so in a low-level language (like Assembly) would lead to unending pain.
This is really the level at which virtual machines start to become common-place.  Java, Scala, C# etc all use a virtual machine to provide an execution environment.  Thus, many mid-level languages don’t compile directly down to the metal (at least, not right away) but represent a blurring between interpreted and compiled languages.  Mid-level languages are almost always defined in terms of low-level languages (e.g. the Java compiler is bootstrapped from C).

Characteristics

  • High level abstractions such as objects (or functionals)
  • Static typing
  • Extremely commonplace (mid-level languages are by far the most widely used)
  • Virtual machines
  • Garbage collection
  • Easy to reason about program flow

High-Level

High-level languages are really interesting if you think about it.  They are essentially mid-level languages which just take the concepts of abstraction and high-level constructs to the extreme.  For example, Java is mostly object-oriented, but it still relies on primitives which are represented directly in memory.  Ruby on the other hand is completely object-oriented.  It has no primitives (outside of the runtime implementation) and everything can be treated as an object.
In short, high-level languages are the logical semantic evolution of mid-level languages.  It makes a lot of sense when you consider the philosophy of simplification and increase of abstraction.  After all, people were n times more productive switching from C to Java with all of its abstractions.  If that really was the case, then can’t we just add more and more layers of abstraction to increase productivity exponentially?
High-level languages tend to be extremely dynamic.  Runtime flow is changed on the fly through the use of things like dynamic typing, open classes, etc.  This sort of technique provides a tremendous amount of flexibility in algorithm design.  However, this sort of mucking about with execution also tends to make the programs harder to reason about.  It can be very difficult to follow the flow of an algorithm written in Ruby.  This “obfuscation of flow” is precisely why I don’t think high-level languages like Ruby are suitable for large applications.  That’s just my opinion though.  

Characteristics

  • Interpreted
  • Dynamic constructs (open classes, message-style methods, etc)
  • Poor performance
  • Concise code
  • Flexible syntax (good for internal DSLs)
  • Hybrid paradigm (object-oriented and functional)
  • Fanatic community
Oddly enough, high-level language developers seem to be much more passionate about their favorite language than low- or mid-level developers.  I’m not entirely sure why it has to be this way, but the trend has been far too universal to ignore (Python, Perl, Ruby, etc).  Ruby is of course the canonical example of this primarily because of the sky-rocket popularity of Rails, but any high-level language has its fanatic evangelists.

Strength of Materials - Strains

3. Strains

Strain is defined a the ratio of change in dimension to original dimension of a body when it is deformed. It is a dimensionless quantity as it is a ratio between two quantities of same dimension.
3.1. Linear Strain

Linear strain of a deformed body is defined as the ratio of the change in length of the body due to the deformation to its original length in the direction of the force. If l is the original length and dl the change in length occurred due to the deformation, the linear strain e induced is given by e=dl/l.
Linear Strain
Linear strain may be a tensile strain, et or a compressive strain ec according as dl refers to an increase in length or a decrease in length of the body. If we consider one of these as +ve then the other should be considered as –ve, as these are opposite in nature.




3.2. Lateral Strain
Lateral strain of a deformed body is defined as the ratio of the change in length (breadth of a rectangular bar or diameter of a circular bar) of the body due to the deformation to its original length (breadth of a rectangular bar or diameter of a circular bar) in the direction perpendicular to the force.
3.3. Volumetric Strain

Volumetric strain of a deformed body is defined as the ratio of the change in volume of the body to the deformation to its original volume. If V is the original volum and dV the change in volume occurred due to the deformation, the volumetric strain ev induced is given by ev =dV/V



Consider a uniform rectangular bar of length l, breadth b and depth d as shown in figure. Its volume V is given by,
Volumetric Strain



This means that volumetric strain of a deformed body is the sum of the linear strains in three mutually perpendicular directions.
 
3.4. Shear Strain

Shear strain is defined as the strain accompanying a shearing action. It is the angle in radian measure through which the body gets distorted when subjected to an external shearing action. It is denoted by *.
Shear Strain



Consider a cube ABCD subjected to equal and opposite forces Q across the top and bottom forces AB and CD. If the bottom face is taken fixed, the cube gets distorted through angle * to the shape ABC’D’. Now strain or deformation per unit length is

Shear strain of cube = CC’ / CD = CC’ / BC = * radian

Thermometer

Definition: A thermometer is a device that measures temperature.  When we touch something we either feel hot or cold or we may not feel anything at all. This relative feeling is a qualitative measurement which can tell whether the body is warm or cold. But this type of observation can not tell how much hot or how much cold it is. Hence, in order to get a quantitative value of temperature we must be able to measure it and get a number corresponding to the degree of hotness or coldness. A thermometer helps us measure this quantity called temperature.
Temperature is that property of a system which can tell whether the system is in thermal equilibrium with another system or not. It is also the degree of hotness or coldness of a body.Thermometric Property: Change in temperature of any material is associated with change in its other properties such as Pressure, Volume, Density, Electrical resistance, color, etc.
These properties which exhibit a known relationship to the change in temperature are known as thermometric properties. Thermometric properties are important because temperature is intangible but the related thermometric properties are tangible in nature. We can not see temperature but we can see its volume and color changing, we can see the pressure rising, we can see it becoming denser or lighter than another reference material. So, we can use the change in volume to correlate the change in temperature and by quantifying the volume change we can quantify temperature. How it works: A thermometer has two basic components, viz. (i) Temperature sensor  (ii) Read out scale The temperature sensor is made up of a material that is capable of showing any change in thermometric property. For example: in the household thermometer generally mercury is used which expands w.r.t. the amount of heating done. The readout scale is obtained as follows. First the thermometric property is let to change over a known temperature range(ice point to steam point). Then the change in thermometric property and the temperature range are correlated so that a relationship can be established. Any unknown temperature can be found out by putting the value of change in the thermometric property in the relationship. Types of thermometers:Thermometers differ by their operating principle and the thermometric property being considered. The following is a list of various types of thermometers used. The thermometric property for each type is given in the bracket.

Platinum resistance thermometer
Platinum resistance thermometer
  • Liquid-in-glass thermometer (Volume/length)
  • Gas thermometer (Volume)
  • Differential expanding bi-metallic thermometer (length)
  • Thermocouple (E.M.F.)
  • Electrical resistance thermometer (Resistance)
  • Optical pyrometer (Color)
  • Difference in density thermometer (Specific gravity)
All the above mentioned types of thermometers are different in terms of operation and construction. But the calibration of each type is similar.
Electrical resistance thermometer:
Qn: A platinum resistance thermometer reads a resistance of 6.4 Ω at 0 °C and 7.8 Ω at 100 °C. Calculate the temperature when the resistance is 7.0 Ω.
Ans: The temperature-resistance relationship for platinum resistance type thermometer is given by
R=Ro (1 + At)
 By putting the values at 0 °C we get
6.4 = Ro (1 + A x 0) = R0 ( 1+ 0) = Ro
⇒ Ro = 6.4 Ω
Now putting the values at 100 °C we get
7.8 = 6.4 ( 1 + A x 100)
⇒ 7.8 / 6.4 = 1 + A x 100
⇒ (7.8 / 6.4) – 1 = A x 100
⇒ A = ((7.8 / 6.4) -1)/100
⇒ A = 0.0021875
For a resistance value of 7.0 Ω we get the relation ship as
7.0 = 6.4 (1 + 0.0021875 x t)
⇒ t = ((7.0/6.4)-1)/0.0021875 = 42.857 °C
Thermocouple:
In a thermocouple two different metals are formed in a junction. One end is embedded in a liquid of known temperature (usually ice) and the other is placed at a point whose temperature is to be measured. The difference in temperature causes a potential difference across the ends as per Seebeck effect. This EMF generated is correlated with the thermal gradient present and calculation is done from the relationship to obtain a value.
Thermoelectric effect or Seebeck effect is the conversion of temperature difference into electric potential difference. This effect is observed in thermoelectric materials when there is difference in temperature at their ends. The energy gradient caused a thermal current to set in through the material and thereby induces an electric potential.
Qn: A thermocouple connected to a millivoltmeter gives EMF at t °C as : ε = 0.36 t – 0.00048 t² mV. If the thermocouple-millivolt meter combine is found to be correctly calibrated at 0 °C and 100 °C, find the temperature measured by the thermocouple when a gas thermometer reads 70 °C.
Ans:
The emf value at 0 °C :
ε = 0.36 x 0 – 0.00048 x 0²
 = 0 mV
The emf value at 100 °C :
ε = 0.36 x 100 – 0.00048 x 100²
 = 31.2 mV
Difference in voltage = (31.2-0) mV
Difference in temperature = (100-0) °C
Rate of change = 100/31.2
EMF read by thermocouple at 70°C is
ε = 0.36 x 70 – 0.00048 x 70²
   = 22.848 mV
Hence, the voltage indicated is
t = (100/31.2) x 22.848 °C
  = 73.23 °C.