Weather Impacted Wars & Migrations in Pre-Recorded History


 

Archaeologists working in the wetlands of Denmark have uncovered 2,000-year-old human remains are revealing that the Germanic “barbarians” were engaging in warfare in northern Europe against other barbarian tribes which had nothing to do with Rome. The research, which was published today in the journal Proceedings of the National Academy of Sciences, provides a unique look at how Germanic tribes memorialized their battles. The sheer magnitude of the number of remains demonstrates that the Germanic armies were clearly organized with leadership. There are no recorded history of this people so all we have to go on are the things left behind.

One thing to emerge is that the climate in Northern Europe was turning very cold. Even by 170AD, when the Roman Emperor Marcus Aurelius was battling the Germanic attempts to invade the south most likely due to climate change. He wrote his Mediataions to keep his mind distracted from the cold at night. He sent his children home to Rome to live with their great-great-aunt Matidia because Marcus thought the evening air of the country was far too cold for them. He even asked his friend for “some particularly eloquent reading matter, something of “your own, or Cato, or Cicero, or Sallust or Gracchus—or some poet, for I need distraction, especially in this kind of way, by reading something that will uplift and diffuse my pressing anxieties.” id/ Ad Antoninum Imperator 4.1 (= Haines 1.300ff), qtd. and tr. Birley, Marcus Aurelius, 120.

It appears from this new evidence that as the weather grew colder and colder, barbarians first fought against each other for resources. It appears that after such inter-tribal warfare proved fruitless, this is when the Germanic invasions of Rome begin. The Germanic invasions of Rome began during 113 BC and lasted until they finally conquered Rome by 596 AD.

Germanic tribes did not produce coins of their own origin. The began manufacturing the first imitations presumably around during the last century BC/first century AD. These were imitations of denarii and quinari (half-denarius) of which a great number appear to have been struck over a wide-ranging territory. Obviously, the barbarian tribes had contact with Rome and imitated their coins, not as counterfeits, but to facilitate trade among themselves. Coins provide evidence of trade as well as political contacts extending into regions with no appreciable recorded history. Here is a Celtic imitation of a gold Stater of Philip II (359-336BC) of Macedonia.

Here we have a barbarian imitation of a barbarian imitation of a gold stater which appears to be a half-stater. This is an imitation made by the Swiss tribe of the Helvetii. Obviously, we have imitations of Greek coinage before Rome was important that included gold.

During the Roman Period, imitative denarii are observed across a large area of the Barbaricum world, in both the Western and the Eastern Germanic territory. We find especially denarii hoards range from Westphalia, Gotland, through Poland (the Przeworsk culture & Wielbark culture) to the Great Hungarian Plain, Moldova, Ukraine and southwestern Russia (Cherniakhiv/Sântana de Mureš cultures).

The great number of imitative coins with identical dies do surface a considerable distance apart suggesting that there was clearly an economic trade and that the imitations were being struck generally from specific centers. One center producing imitations appears to be of the Gothic region which was located in Ukraine (Cherniakhiv culture). There is an extensive concentration of imitative denarii recorded in this region and further east. The workmanship is reasonable, but still easily distinguished from genuine coins.

The early Celtic tribes generally copied Greek coins, but the iconography of imitative Greek staters appear more like abstract art than faithful designs attempting to actually imitate the Greek staters. In Ukraine, some imitative denarii were made of soft alloy, cast in bronze molds.

Some imitative denarii produced in Ukraine were pierced as was the case often in India where the coins would be strung together which was clearly a Chinese tradition. The Ukrainian imitations are often lighter and slightly smaller as well.

What is clear is that as the weather turned colder, the decline and fall of the Roman Empire began from about 180AD. This also appears to result in more frequent Germanic invasions from the North no doubt as weather reduced the food supply. A similar result appears between the Greek Heroic period and its Hellenistic period where Greece went into a Dark Age. This corresponds to the invasions of Northern Africa by the famous “Sea People” who were probably Greeks. The Philistine culture has revealed Maeceanisn pottery suggesting that they may have been in contact with the Sea People or may have merged with them. Recorded history is vague and there are indications of revolts even against Egypt by the Canaanites.

Climate Change & the 43-Year Cycle


QUESTION: Dear Martin,

I was browsing through the website of NASA and came across a graph (see attachment).
It shows the temperature differences between 1880 and 2012. As far as I can see there seems to be a +/- 40-year cycle. I have edited it with the red vertical lines. The end of the current 40-year cycle should be around 2020, which, if I am not mistaken, coincides with a date you have mentioned previously in one of your blogs. This could mean that around that date a new 40-year cycle should start and temperatures should decline. What is your opinion on this?

Best regards,

SG

ANSWER: Yes, no bad. Look closely. The Japanese data seems to be the least messed with. The others have recalculated the data numerous times to get the biggest effect. Our model shows a 43-year cycle to be precise. That is driven by the energy output of the sun for it tests back into ancient times as well. The Japanese data is the least “politically correct” data whereas the US versions have an agenda behind them to justify raising taxes as if we can really change the climate of the entire planet. They get away with this because they confuse people with pollution and climate change. Hospital waste dumped in the sea does not change the climate but is disgusting and dangerous in spreading disease.

The energy output of the sun peaked in 2015 and has been crashing faster than anyone thought possible. The fascinating thing as you look at the chart and see the cycle. They do not. They forecast linearly whatever trend is in motion will last forever.

Analysis of Global Temperature Trends, April, 2018, what’s really going on with the Climate?


The analysis and plots shown here are based on the following two data series. First NASA-GISS estimates of a global temperature shown as an anomaly (converted to degrees Celsius) as shown in their table Land Ocean Temperature Index (LOTI) and shown in Chart 1 as the red plot labeled NASA the scale for the temperatures is on the left. The NASA LOTI temperatures are shown as a 12 month moving average because of the large monthly variation. Second NOAA-ESRL Carbon Dioxide (CO2) values in Parts Per Million (PPM) which are shown in Chart 1 as a black plot labeled NOAA the scale for CO2 is shown on the right. However if you don’t have time to read this entire report go to Chart 7 and it will show you exactly what is going on with CO2 and global temperatures; I guarantee that this chart is based on NASA data and is 100% accurate.

NASA published data as stated in the first paragraph is shown as an anomaly, but what is a temperature anomaly?  An anomaly is a deviation from some base value normally an average that is fixed. There were two problems with the system that NASA picked which were number one there is no “actual” global temperature and two since climate is a variable there cannot be a real base to measure from. NASA known for its science and engineering expertise back in the day thought it could get around these issues and created a system to do so. First they developed a computer model which took readings from all over the planet and made required adjustments to them which they called homogenization and came up with the estimated global temperature. Second they picked the period 1950 to 1980 (30 years) and averaged the values found in that period and came up with 14.00 degrees Celsius and make that their base.  Then they took the calculated monthly temperature and subtracted the base from it which gave them the anomaly. The problem is that both are arbitrary.

Now that we have a base to work with we are going to add to Chart 1 three things. The first is a trend line of the growth in CO2 since that is according to the government through NASA and NOAA the entire basis for climate change. That plot is superimposed over the black plot of the actual NOAA CO2 values as the cyan line labeled as the CO2 Model and one can see there is a very good fit to the actual NOAA values so there should be no dispute about its validity, and it’s historically accurate.  This plot allows us to make projections to future global temperatures according to the projected level of CO2 .  The second added item is James E. Hansen’s 1988 Scenario B data, which is the very core of the IPCC Global Climate models (GCM’s) and which was based on a CO2 sensitivity value of 3.0O Celsius per doubling of CO2. This plot is shown here in lavender and is part of a presentation that Hansen showed to congress in 1988 when the UN was about to set up the International Panel on Climate Change (IPCC) and this plot is labeled as Hansen Scenario B which Hansen stated was the most likely to happen based on his 1979 climate theories’.  The third item is the current plot of the most likely temperature of the planet based on the growth of CO2 published by the IPCC. This plot is shown in Red and is labeled as IPCC AR5 A2 as that is the table where the data was found. This plot is a GCM computer projection of the planets temperature based on the complex relationships developed on the levels of CO2 by the IPCC primarily though NASS and NOAA.

It can be seen in Chart 2 that the lavender plot and the Hansen plot are very close from 1965 to around 2000 after that, from 2000 to 2014, there is a very large and deviation reaching close to .5 degrees Celsius in 2015, which is not an insubstantial number.  Also of note is that there doesn’t seem to be a good correlation between the growth in CO2 and the increase in the planets temperature. The CO2 is going up in a log function and the Temperature was going down until 2015 and then there was a mysterious spike up. That unexplained change in temperature direction appeared to have occurred between 2013 and 2014 and is the subject of this monthly paper.

Next we have Chart 3 which is developed from the raw data from NASS and NOAA as shown in Chart 1.  This plot was made first by adding ten years blocks of temperature and CO2 as indicated in the Chart 1 and diving by 120 to give an average for each.  Then the average Temperature was divided by the average CO2 to give degrees of temperature increase per PPM of CO2. After that was plotted it appeared that there were two different curves. The first was from block 1965-1974 through block 2004-2014 shown as Black Dots and the second was from block 1995-2004 through block 2005-2017 shown as Black Dashes. When trend lines were added they were both almost perfect fits to the raw data and so you cannot see the data points very well on Chart 2.  These blocks were picked to represent the entire period of time where we had both NASA temperature data and NOAA CO2 levels.

On Chart 3 there are two sets of color coded information. The first is Cyan plot and the Cyan box with the equation in it along with the R2 value of 1.0 are for the first series from block 1965-1974 through block 2004-2014. The other is the Red plot and the Red box with the equation in it along with the R2 value of 1.0 which are for the first series from block 1965-1974 through block 2004-2017. We can speculate on how this change happened but it can’t be said that the plot change is not real; however additional data will be required to actually prove that something has changed.

In summary the Cyan data set indicates a diminishing effect of CO2 on global temperature for about 54 years and the Red data set represents an increasing effect of CO2 on global temperature for the past 3 years. Since both data sets have an R2 value of 1.00 the trend lines cannot be in question.

Continuing the analysis of what happened to the NASA data in table LOTI from Chart 3, the following Chart 4 was constructed from the same NASA data. It’s very sad to say but it seems to prove without much doubt that the global temperatures have been manipulated by NASA probably at the request of the federal government such that a case could be made for supporting the COP21 Paris climate conference in December 2015 by showing that the earth was much hotter than it actually was. The dates on the x axis are the date of the NASA LOTI download file. The plots for specific date groupings are set such that one can see what that date range did in each separate NASA download. The proof is shown in Chart 4 below and a discussion will follow below Chart 4 on how Chart 4 was constructed.

At the bottom of Chart 4 is a blue trend line of NASA LOTI temperatures prior to 1950 and starting in2012 the values started going down, getting colder. At the same time the NASA LOTI temperatures from 2012 to the present went up as shown in the red line.  There was no change in the base period, black line. This cannot happen with random variables they will cancel each other out; this could only be caused by specific program changes in the process that NASA and NOAA use, in other words it is intentional. So there can be no other reason but an attempt to support the adoption of the Climate accord agreement by the administration, and they were successful as it was agreed to in Paris at COP21.

How this table was constructed is important so a discussion is needed. As stated in the opening paragraph of this paper NASA publishes a table of the estimated global temperature each month as anomalies from a base of 14 degrees Celsius. This table starts with January 1880 and runs to the current date. The new table typical comes out mid-month with almost 1,700 values for the previous month. The process that is used to create this Table is very complex and is called homogenization. What that means is that the entire table is recreated each month and what that also means is that the temperature value for any given month is a variable.

When I realized the extent of that in 2012 I started to save the printouts of the NASA LOTI tables and I went back and found a few of them from when I started this project in 2007. When I started this project what I did is type in all the values from the NASA table into a spreadsheet each month which was a daunting task and I was very happy when NASA started to publish a csv file along with the text of the LOTI data. Then all I had to do is create a routine in excel that would turn the table format into a column format.  There are now 68 months in the spreadsheet, when I started this method in 2012 there were maybe only a dozen. The values are residing in the spreadsheet as columns going from left to right so that the individual months are lined up side by side. This makes comparison of months very easy. One note is required here, when I started this model in 07 and for several years thereafter all I was doing is adding the current NASA LOTI current months number to the existing file, a single column, and it never occurred to me that the prior numbers were changing. The past was fixed, so I thought. This was also the way I was entering the NOAA CO2 data which doesn’t change over time.

The original goal was to see if the changes were just random or rounding errors. If that was so then they would wash out over time especially if I grouped the monthly data into blocks. I’ve used both 10 year (120 values) and 20 year (240 values) blocks which would be enough to maintain a fixed number if it was random or rounding. What I found was something quite different after I had a dozen or so columns in the spreadsheet, it appeared that NASA was making the past colder and the present warmer. And the purpose of the previous two Charts 3 and 4 is to show the result. Chart 4 is a bit complex but I have not found a better way to show what happened.

From 1880 to 1960 I used four 20 year blocks.  Then I needed the base so there is a 30 year block from 1950 to 1980 and lastly four 10 year blocks from 1980 to the present. The last block is not yet complete as it will run to December 2019. Because the 30 year base block is fixed at 14.0 degrees Celsius there wasn’t much point in charting those individual yearly values even though there was some minor movement in those numbers. That raises an interesting issue for how can the base numbers not change and all the other numbers from 1880 to 2017 can change each month? A note, for each data set of years the plot on Chart 4 should be a straight line from left to right; very minor fluctuation would be OK. For example the plot for 1930 to 1949 (hidden behind the black plot) is what would be normally expected. This is the only plot that doesn’t show major manipulation.

In the four data sets in the 1880 to 1940 blocks in Chart 4 all have moved down probably about a .25 degree Celsius which is not insignificant. So the bottom line is that NASA made all the values from 1880 to 1940 colder by an average of a quarter of a degree Celsius. So that alone accounts for a high percentage of the supposed global warming that NASA shows. From 1980 to 2009 the data change appears to add another .1 degrees Celsius making the apparent differential between data from early 00’s to the present about .35 degrees greater than it was before 2009. That is not random that is a major change and clearly shows manipulation. I would probably never had caught this is if I hadn’t put the values in column format. Looking at all the data from 2008 to 2014 we find that around 2008 NASA showed that the planet had warmed about .75 degrees, Blue double arrow, from the 19th century. Then in 2014, four years later NASA showed that the planet had warmed about .95 degrees Red double arrow from the 19th century. However it gets a worse after that.

The change started in 2012, Green Oval, and Global temperature jumped almost a quarter of a degree by December 2015 just as the COP21 conference was in session. The temperatures kept going up with an eventual increase in global temperature of about 1.2 degrees Celsius in late 2016. At that point with the pressure off NASA appears to be erasing what they did as the global temperatures have now started back down.  I’m not sure how many know of this blatant manipulation but it is serious. This is not science.

Now we need to consider other factors than CO2 on Climate change.  The fault that occurred in the work that was done in the 1980’s was in assuming that there was an optimum or constant global temperature and therefore any change that was being observed was from the increasing amount of CO2 in the atmosphere.  There may have been correlation but it was never proved that there was causation (high R2 value) between CO2 and global temperatures; Chart 3 clearly shows there is not. With that assumption, which limited options, we moved from true science into the realm of political science.  True science has an open mind and finds relationships that work in matching observations with predictions.  Political science changes history and/or facts to match the desires of the politicians. Since the politicians control the money political science is what we get; which means that what we get may not be technically correct.

A decade ago when I started looking at “climate” change the first thing I did was look at geological temperature changes since it is well known that the climate is not a constant; I learned that 53 years ago in my undergrad geology and climatology courses in 1964. The next paragraph explains currently observed patterns in climate related to this subject and is historical accurate.

Ignoring the last Ice Age which ended some 11,000 years ago when a good portion of the Northern hemisphere was under miles of ice the following observations give a starting point to any serious study on the subject of climate. First, there is a clear up and down movement in global temperatures with a 1,000 some year cycle going back at least 3,000 to 4,000 years; probably because of the apsidal precession of the earth’s orbit of about 20,000 years for a complete cycle. However about every 10,000 years the seasons are reversed making the winter colder and the summer warmer in the northern hemisphere. 10,000 years from now the seasons will be reversed again. Secondly, there are also 60 to 70 year cycles in the Pacific and the Atlantic oceans that are well documented. These are known as the Atlantic Multi Decadal Oscillations (AMO) in the Atlantic and as La Nina and El Nino in the Pacific. Thirdly, we also know that there are greenhouse gases such as carbon dioxide that can affect global temperatures. Lastly the National Academy of Sciences (NAS) estimated that carbon dioxide had a doubling rate of 3.0O Celsius plus or minus 1.5O Celsius in 1979 when there were only two studies available and one for sure and maybe both were not peer reviewed.

The result of looking objectively at the three possible sources of global temperature changes was a series of equations based on these observations that when added together produced a sinusoidal curve that seemed to follow NASA published temperatures very closely when first developed in 2007, and modified a few years later when it was found the short and long cycles were related to multiples of Pi.  Since this curve was based on observed temperature patterns it was called a Pattern Climate Model (PCM) which has been described in previous papers and posts on my blog and since it is generated by “equations” many assume it is some form of least squares curve fitting, which it is not. It does seem to be related to ocean currents where the bulk of the planet’s surface heat is stored.

Chart 5 shows the PCM a composite of two cycles and CO2. There is a long trend, 1036.7 years with an up and down of 1.65O Celsius (.00396O C per year) we in the up portion of that trend. Then  there is a 69.1 year cycle that moves the trend line up and then down a total of 0.29O Celsius and we are now in the downward portion of that trend (-.01491O C per year), which will continue until around ~2035. Lastly, there is CO2 currently adding about .0079O Celsius per year so together they all basically wash out at -.0039O C per year, which matches the current holding pattern we were experiencing until 2014. After about 2035 the short cycle will have bottomed and turn up and all three will be on the upswing again duplicating what was observed in the 1980’s.  Note: the values shown here are only representative from what is in the model.

When using a 12 month running average for global temperatures up until 2014 the PCM model was within +/- .01 degrees of what NASA was publishing in their LOTI table since the early 1960’s as shown in Chart 5. Further the back projection of the PCM plot matched historical records and global temperatures going back past the time of Christ. It should also be considered that geologically CO2 levels have reached levels many times that of the current 400 ppm without destroying the planet so the current hysteria over the current very small numbers can only be explained by political science not real science.

The nest step in this analysis is to put all of the known data and projections into Chart 6 which contains: NASA’s temperatures plot, NOAA’s CO2 plot, the CO2 model plot, the PCM model plot, Hansen’s Scenario B plot, and lastly the IPCC AR5 A2 global temperature plot. With that done we can look at the results and try to make some sense of what is going on with the various arms of the federal government that are promoting that we tax carbon based fuels to eliminate them since they are responsible for the global temperature level  going up.  As previously stated when the government pours money into the sciences the sciences respond with technical papers the support the governments views, this is what I call political science verses real science as was done prior to the 1980’s; money talks and BS walks as everyone on the street knows.

Chart 6 shows a good overview and contains no data manipulation and the only change that was made was to convert the NASA anomalies back to degrees Celsius to make it more readable to lay people.  This is only a change in units and has no bearing on the look.  We also need to understand the NASA homogenization process and its relationship to the 30 year base period. The portion in the black circle contains the NASA base period of 14.00 degrees Celsius and the reason it’s brought up here is that the Homogenization process causes the global temperatures to move around since the entire data base all the way back to 1880 is recalculated each month.  But since the base has to stay at 14.00 degrees Celsius the program must be set to not allow changes in that period of time. I’m sure the programmers have fun with that. Prior work here has shown how this creates a teeter totter effect with the data plots, some of which have recently been significant.

Next Chart 7 looks at the period from 2010 to 2020 so we can see where a change in CO2 of only a few ppm has caused a major change in the global temperature way beyond anything previously shown in any published NASA data. There are two black ovals on Chart 7 one at the top of Chart 7 which is a black oval around the CO2 levels from 2012 to 2016 and part of 2017 and it’s very obvious that there has been very little change, maybe 7 ppm or about 1.9%. Then at the bottom of Chart 7 is another black oval around the NASA global temperature levels for the same period and its very obvious that there has been a large change, almost .50 degrees Celsius or about 3.1%. There has never been such a large increase in temperature from such a small increase in CO2. By contrast the previous comparable period of the last part of 2010 through 2013 shows about the same increase for CO2 at 1.1% but no increase for global temperature but actually small decrease.

Clarification is needed here as the plot seems to show the jump in temperature in 2016 not 2015; this is a result of the large jump in temperature shown by NASA. Since we are using a 12 month moving average and the increase occurred in only a few months it actually shifted the curve into 2016. The raw data for December 2015 showed the temperature at 15.12 degrees Celsius compared to December 2014 where it was 14.78 degrees Celsius. The actual peak was in February 2016 at 15.35 degrees Celsius.   With the global temperature over 15.0 Celsius at COP21 the climate accord was approved and the manipulation was a success. After COP21 the need for Fake Warming was no longer needed and so we are now seeing a downward trend developing.

In summary, the IPCC models were designed before a true picture of the world’s climate was understood. During the 1980’s and 1990’s CO2 levels were going up and the world temperature was also going up so there appeared to be correlation and causation. The mistake that was made was looking at only a ~20 year period when the real variations in climate all move in much longer cycles of decades and centuries.  Those other cycles can be observed in the NASA data but they were ignored for some reason.  By ignoring those actual geological trends and focusing only on CO2 the Global Climate Models will be unable to correctly plot global temperatures until they are fixed. Also the temperature data from 1850 to 1880 was dropped for some reason as it showed a lower temperature that supported the PCM cycle shown in this paper.

In summary we have Chart 8 which shows why CO2 is not increasing the temperature of the planet by any meaningful amount. The problem, intentional or not, goes back to physics and how we show information. It’s critical that when we talk to non scientists that information is properly displayed. And nowhere is this more important than when we are discussing temperature.  When we talk about weather and local temperatures its going be in Celsius (C) in the EU or degrees Fahrenheit (F) in America e.g. for the base temperature that NASA uses it’s 14.00 C or 57.20 F; but these are both relative measures and do not tell us how much heat (thermal energy) is there. To know that we must use Kelvin (K) and that would be 287.150 K and all three of those numbers 14.00 C, 57.20 F, and 287.150 K are exactly the same temperature, just using a different base. But if the current temperature is 15.00 C that is a 7.1% increase in C, a 3.1% increase in F and a .35% increase in K; so which one is real? The answer is .35% because Kelvin is the only one that measures the total energy!

To show this graphically Chart 8 was constructed by plotting CO2 as a percentage increase from when it was first measured in 1958 the Black plot, the scale is on the left and it shows CO2 going up about 28.5% by February of 2018. That is a large change as anyone would agree.  Now how about temperature, well when we look at the percentage change in temperature using the proper units Kelvin we find that the changes in global temperature are almost un-measurable. The red plot, also starting in 1958, shows that the thermal energy in the earth’s atmosphere has varied by less than +/- .17%; while CO2 has increased by 28.3% which is over 80 times that of increase in temperature. So is there really a problem here?

Lastly, Chart 9 shows what a plot of the PCM model, in yellow, would look like from the year 1400 to the year 2900. This plot matches reasonably well with recorded history and fits the current NASA-GISS table LOTI data, in red, very closely, despite homogenization.  I do understand that this PCM model is not based on physics but it is also not some statistical curve fitting. It’s based on observed reoccurring patterns in the climate. These patterns can be modeled and when they are, you get a plot that works better than any of the IPCC’s GCM’s. If the real conditions that create these patterns do not change and CO2 continues to increase to 800 ppm or even 1000 ppm then this model will work well into the foreseeable future.  150 years from now global temperatures will peak at around 15.750 to 16.000 C and then will be on the downside of the long cycle for the next ~500 years.

The overall effect of CO2 reaching levels of 1000 ppm or even higher will be about 1.50 C which is about the same as that of the long cycle.  The Green plot on Chart 9 shows the observed pattern with no change in CO2 from the pre-industrial era of ~280 ppm. CO2 cannot affect global temperatures more than 1.500 C +/- no matter what the ppm level of CO2 is. The reason being that the CO2 sensitivity value is not 3.00 per doubling of CO2 but less than 1.00 C per doubling of CO2 as shown in more current scientific work and it’s a logistics curve not a log curve.

The purpose of this post is to make people aware of the errors inherent in the IPCC models so that they can be corrected. 

The Obama administration’s “need” for a binding UN climate treaty with mandated CO2 reductions in Europe and America was achieved as predicted at the COP12 conference in Paris in December 2015. To support this endeavor NASA was forced to show ever increasing global temperatures that will make less and less sense based on observations and satellite data which will all be dismissed or ignored.  Within a few years the manipulation will be obvious even to those without knowledge in the subject, but by then it will be to late the damage to the reputation of science will have been done.

In closing keep this in mind. The current panic generated by the government using political science is that the current global temperature of around 15.0O Celsius is an increase of 7.14% from the 1960’s when the global temperature was 14.0O Celsius; and that does seem like a lot. However those views would be in error as the actual increase in thermal energy, as measured by temperature, would be only .35% because we must use Kelvin not Celsius when working with heat energy. When we use kelvin the temperature goes from 287.15O K to 288.15O K which is only .35% not 7.14% about 1/20 of what is implied by the IPCC. What the IPCC shows is not technically wrong as much as it is extremely misleading to anyone without a very strong science background.

Sir Karl Raimund Popper (28 July 1902 – 17 September 1994) was an Austrian and British philosopher and a professor at the London School of Economics. He is considered one of the most influential philosophers for science of the 20th century, and he also wrote extensively on social and political philosophy. The following quotes of his apply to this subject.

If we are uncritical we shall always find what we want: we shall look for, and find, confirmations, and we shall look away from, and not see, whatever might be dangerous to our pet theories.

Whenever a theory appears to you as the only possible one, take this as a sign that you have neither understood the theory nor the problem which it was intended to solve.

… (S)cience is one of the very few human activities — perhaps the only one — in which errors are systematically criticized and fairly often, in time, corrected.

Part I – The FISA Court Grants The Authority, Not The Ability…


There is a meeting scheduled tomorrow between key congressional oversight committee heads (Nunes, Gowdy etc.) and leadership of the FBI (Director Wray), DOJ (Edward O’Callaghan) and ODNI (Dan Coats).  The meeting was set up by White House Chief of Staff John Kelly, and the purpose of the meeting is to come to some agreement on access to documents being withheld by the DOJ, DOJ-NSD and FBI.

However, amid the ongoing debate over spies and informants used by the CIA and FBI to conduct political surveillance, there’s an aspect of the ongoing investigation that seems to be entirely overlooked.

On January 7th, 2016 the Inspector General of the National Security Agency, George Ellard, submitted a mandatory compliance report outlining the status of the NSA’s ability to monitor the access of users within the NSA database.

I’m including the full report below in pdf format because it is important to understand what the NSA inspector general said.  I strongly urge you to read it (despite the redactions) because the larger issues remain visible within the report:

https://www.scribd.com/embeds/379947972/content?start_page=1&view_mode=&access_key=key-TIME5AxQlMcwsmxN5D4F

.

For those who don’t read it, a quick summary would be: We don’t actually have control over how this massive amount of meta-data is being used, searched, queried, and how the information within it is being filtered, minimized or extracted.

The NSA is essentially admitting to being overwhelmed with the scale of the data being collected, and they are similarly overwhelmed by the number of people who have access to this data. Of course the NSA Inspector General uses use much more techno-terminology like:

“Agency controls for monitoring query compliance have not been completely developed.”

…And: “The Agency has no process to reliably identify queries performed using selectors associated with 704 and 705(b) targets.”

…And more alarmingly: “We identified another [redacted] queries that were performed outside the targeting authorization periods in E.O. 12333 data, which is prohibited by the E.O. 12333 minimization procedures. We also identified queries performed using USP selectors in FAA §702 upstream data, which is prohibited by the FAA §702 minimization procedures.”

In short: There’s a ton of activity that we know is taking place that is not supposed to be happening; and we really don’t know the full scale of the abuse we know to be happening -but can’t get our arms around- because, well, there’s just so damn much of it.

This is a critical part of the Obama-term spying and surveillance issue that is being missed by current media reporting.

Given the highly political nature of the intelligence activity we are seeing lately, who do you think would be exploiting those system-wide database vulnerabilities, and why?

Yep, likely those carrying the most entrenched political agenda and ideology. Vested interests both inside the government (career deep state agency staff writ large), and outside government (all those pesky contractors we keep hearing about).

The FBI and NSA database is essentially a vault of unlimited information; the mother-lode of metadata and electronic record-keeping of everything.  Essentially, the holy grail for political operatives who would like to exploit and leverage the information.

Do you want to conduct opposition research?  Those databases would contain such exhaustive amounts of information you could assign a dozen users per target and still not scratch the surface of what is available.  Bank records, purchases, social media accounts, phone records, GPS locations, travel patterns, vehicle locations, home/work ip addresses, emails, audio-activation, pictures, CCTV captures, the works.

Through electronic surveillance you could spy and surveil anyone; in almost real time if focused, motivated and intended. What stands between this massive surveillance system being used/abused for those political purposes or not?

Nothing.

Really.

Unless you want to define self-restraint and the morality of the person who has access as an impediment.  Extractions or leaks are only unlawful if you are on the wrong ideological side of those who control the political reigns of power…. Just ask James Comey.  If you didn’t intend to break the law, you didn’t.

Do you think such a network could/would be exploited or weaponized for political purposes?   Well, did you think the IRS would be weaponized in 2010?  Think about it…

Now, consider yourself a person motivated by nefarious political intent.  Or perhaps like Lisa Page and Peter Strzok you are motivated by a higher calling to protect our country from outcomes of an election you don’t agree with.  Perhaps you are authorized by willful blindness of those who are suppose to be the gatekeepers. Now, what does going to the FISA Court gain you, that you don’t already have?

Going to the FISA Court gains you “authority“; it doesn’t gain you access.  You already have the access.

There’s a saying: “It’s easier to ask for forgiveness, than permission.”

As you think of that phrase, consider the April 2017 declassified FISA Court ruling we often reference, along with the admissions made to the FISA court by the DOJ, FBI and NSA in October of 2016.  Approximately 85% of all FISA-702(17) “about” queries were unlawful violations.

What was the DOJ, FBI and NSA asking the FISA Court for: ‘forgiveness’, or ‘permission‘?

I don’t want to overwhelm any single discussion article with too much cited information.  It can become overwhelming to try and keep track of it all.  However, I would strongly suggest that based on the intelligence communities own admissions, throughout 2015 and the beginning of 2016 – all of the non-preferred presidential candidates were under electronic surveillance by people in and outside of government who had access to this holy grail of opposition research.

It wasn’t until NSA Director Mike Rogers shut down contractor access to the system in April 2016 that political exploitation of the FBI and NSA databases was impeded.  Not stopped entirely, simply impeded.

(FISA 99-page Opinion – Rosemary Collyer Presiding Judge)

Remember this name: John P Carlin.  It’s going to become much more important as the days and weeks progress.  He plays a key role in part of the activity in 2016 and connects the story of Carter Page to the FBI and CIA surveillance operation. {Preview Here}

On September 26th, 2016, the head of the Department of Justice National Security Division, John P Carlin (pictured above) filed the required certification letter (full pdf below) with the FISA court for the year 2016.

Mr. Carlin DID NOT include the Inspector General report from January 2016; and Carlin did not notify the FISA court of the compliance audit requested by NSA Director Mike Rogers as an outcome of that IG report (March 2016).

https://www.scribd.com/embeds/379950457/content?start_page=1&view_mode=&access_key=key-PEblHg2u9BNmsecVyO7W

.

The next day, September 27th, 2016, John Carlin announced his resignation.

From an excellent MarketWatch Timeline:

♦On October 4, 2016, a standard follow-up hearing on the 2016 Section 702 Certification was held (Page 19). Carlin was present at the hearing. Again, Carlin made no disclosure of FISA Abuse. This would be noted by the Court later (see below).

♦On October 15, 2016, Carlin formally left the NSD.

♦On October 20 2016, Rogers was briefed by the NSA compliance officer on findings from the 702 NSA compliance audit. The audit had uncovered numerous “About” Query violations (Senate testimony).

♦On October 21, 2016, Rogers shut down all “About Query” activity. He reported his findings to the DOJ (Senate testimony & inferences from Court Ruling).

♦On October 21 2016, the DOJ & FBI sought and received a Title I FISA probable cause order authorizing electronic surveillance on Carter Page from the FISA Court. At this point, the FISA Court was still unaware of the Section 702 violations.

♦On October 24, 2016, Rogers verbally informed the FISA Court of his findings (Page 4 of Court Ruling).

♦On October 26, 2016, Rogers appeared formally before the FISA Court and presented the written findings of his audit (Page 4, 14 & 19 of Court Ruling & Senate testimony).

The outcome of the October 26th formal filing by NSA Director Mike Rogers was the 99-page ruling from presiding FISA Judge Rosemary Collyer:

https://www.scribd.com/embeds/349542716/content?start_page=1&view_mode=&access_key=key-72P5FzpI44KMOuOPZrt1

FISA Court Presiding Judge Rosemary Collyer; Primary Intelligence Oversight Chairman Devin Nunes; Primary Judiciary Oversight Chairman Bob Goodlatte.

I would suggest to everyone following this deeply complex spying and surveillance story, that it was AFTER Mike Rogers initially began blocking use of the databases for unauthorized searches (March 2016), that those political operatives (outside contractors) shifted their spying and surveillance activity.

Following this timeline it was after March 2016 when the use of human assets, CIA and FBI agents, became more of a necessary function within the process.

After March and April of 2016 is where George Papadopoulos, Carter Page, and Stefan Halper become needed by the intelligence apparatus. It was in April 2016 when Fusion-GPS, Nellie Ohr and Chris Steele became more important.

I’ll get to the details of how each of them engage within the larger surveillance operation with greater specificity and citation in Part II.  And I’m pretty sure I can prove this:

.

President Trump Keynote Speech – Susan B Anthony Gala for Life


President Trump delivers the keynote speech at the 11th Annual Susan B Anthony Gala for Life:

Everything Michael Caputo Said in Last 24 Hours Was One Big Misunderstanding….


Former campaign aide Michael Caputo was on Fox News Monday Night and Tuesday afternoon with some rather dramatic claims about being targeted for exploitation by a nefarious intelligence community operative.  Apparently it was all just a big misunderstanding.

Misunderstanding Story – EXPLAINED HERE

Whoopsie daisy.

Surveillance Target Sam Clovis Discusses His Contacts With FBI/CIA Intelligence Agent Stefan Halper…


I’m working on a rather comprehensive outline to put all of the actions within the larger political operation together; things in sequence will make much more sense. However, in the interim, here’s former Trump campaign adviser Sam Clovis discussing his encounters with the CIA controlled intelligence agent Stefan Halper.

[*Note* For the sake of context, intellectual honesty and narrative transparency, it is important to note that Sam Clovis is legally represented by Victoria Toensing, wife of Joe diGenova. The reason for this understanding will become more obvious at a later date.] WATCH:

.

Note how the thread/relationship connecting Stefan Halper and Carter Page is surfacing with growing clarity.

It has been admitted that Stefan Halper was an asset paid by the U.S. government, contracted by the CIA and FBI, for activity within FBI operation “Crossfire Hurricane”. I would draw more specific attention to how often we are hearing the word “contractor” in relationship to these intelligence operations.

Keep your eyes and ears open for the term “contractor“. The frequency of that term within these stories is not an accidental.

Edward Snowden was an NSA contractor.  Reality Winner was an NSA contractor.  Christopher Steele was an FBI contractor. Daniel Richman was an FBI contractor. Nellie Ohr was a CIA contractor. Stefan Halper was an FBI/CIA Contractor. Glenn Simpson was an FBI/CIA contractor.  Fusion-GPS was an FBI contractor.  Crowdstrike is an FBI contractor. The FBI approached Oleg Deripaska about being a repeat FBI contractor etc. ….Within the larger story, pay close attention to where/when you see the term “contractor“.

The FISA court has become a misnomer in this storyline.  The metadata, the raw material needed for electronic surveillance and spy operations, always exists regardless of engagement with the FISA court.

The extraction, review and analysis of that metadata does not require a FISA court ordered warrant; it only requires a person have “access” to do the search.  Nothing more.  “Access” is the key to electronic surveillance and data-mining, not “authority”.

Remember that.

Access is the key.

Hence, contractor access opens doors.

The ideology of the contractor determines what they do with that access.

.

https://www.scribd.com/embeds/349542716/content?start_page=1&view_mode=&access_key=key-72P5FzpI44KMOuOPZrt1

.

Matt Gaetz Discusses Call for Independent Probe of DOJ and FBI Misconduct…


Following on the heels of the House of Representatives filing a resolution today outlining evidence of FBI and DOJ political corruption – and calling for a second special counsel, Representative Matt Gaetz appears on CNN to discuss the granular issues.

In this interview democrat operative Jake Tapper attempts to lay multiple traps for Matt Gaetz to walk into.  However, Matt Gaetz uses a firm understanding of the facts to avoid the narrative efforts of Tapper and deconstructs the nonsense with ninja-level retort.

House of Representatives, 19 Members, Outline Resolution of FBI and DOJ Misconduct – Request Second Special Counsel…


Earlier today nineteen members of the U.S. House of Representatives filed a resolution outlining widespread corruption within the institutions of the FBI and U.S. Department of Justice.  The 12-page resolution (full pdf below) highlights examples of known politicization by the FBI and DOJ and calls for the appointment of a second special counsel.

With the introduction of the resolution the house members including: Reps. Lee Zeldin (R-NY), Mark Meadows (R-NC), Jim Jordan (R-OH), Ron DeSantis (R-FL), and Matt Gaetz (R-FL) held a press availability to detail the specific concerns that lead to their request for a second special counsel. Watch:

Here’s the resolution outlining the reasoning and purpose:

https://www.scribd.com/embeds/379917465/content?start_page=1&view_mode=&access_key=key-xYtJApDeWiFT2E1Tz8eg

(PDF House Link)

.

Contemporary Sources



(from the collection of Martin Armstrong)

QUESTION: A friend of mine has one of your original Greatest Bull Market in History books. I noticed that you quote your direct sources from newspapers and magazines rather than other people histories. Did you find the newspaper a better source of truth?

HY

ANSWER: Absolutely. I have collected newspapers from the 1600s to date. They are an invaluable contemporary source of what was being talked about and events at that moment in time. Here is a rare early publication known as The History of Popery, printed in London October 3rd, 1679. This newspaper was an example of the religious issues in Britain that led to the Civil War. This was an anti-Catholic newspaper, which is rather rare for it had a short life before it was suppressed. This one was on Purgatory stating that this is “[t]he belief thereof a Popish Article of Faith.” It goes into the brewing hatred of Catholic in England at that moment in time. You cannot get a sense of real beliefs by reading someone’s interpretation. They tend to pick and choose events or comments to support a predetermined conclusion.

Take the Great Crash which has been the seminal foundation of the history of the causes of the Great Depressions. It is a total piece of socialist propaganda in my opinion. Nowhere in this book do you even find a mention of the Sovereign Debt Crisis. The predetermined conclusion was to be corporate greed and that is why we needed big government. The same is true of the Creature of Jekyll Island and countless other books. When I read Herbert Hoover’s Memoirs, and all the documentation he put in there with the letters between the heads of state, there was NOT a single book written about the Great Depression that was fair, honest and approached the subject with a quest to followed the facts to see where they truly led.

I studied ancient languages. So I learned to read Latin and ancient Greek. This gave me an unusual background when I was interested in economics. Adam Smith’s teacher was Francis Hutcheson (1694–1746) who was a professor of Philosophy at the University of Glasgow. In 1742, Hutcheson published books entitled Short Introduction to Moral Philosophy, Elements of the Law of Nature, and the third one carried the title of The Principle of Oeconomics and Politics. The latter is the book that resurrected “Oikonomikos” of the Athenian Xenophon’s work from its Greek origin during the 4th century BC giving us the word ECONOMICS (oikos” meaning household and the complex root “nem” meaning “to regulate, to administer, to control”).

To my complete shock, Hutcheson translated this into Latin as “Oeconomics” and later into English giving us “Economics” but he also followed the content of Xenophon’s work chapter by chapter. Then the very order of topics discussed in the economic portion of Hutcheson’s System of Moral Philosophy, published in1755, was repeated by Smith in his Glasgow Lectures and again in the Wealth of Nations. So it is interesting to me that the source was really Xenophon, who received no credit.

Over the years, I have bought up newspapers from libraries as they close or get rid of old materials. In the good old days, they would bound the newspaper for each year. Here is the Chicago Daily Tribune for the year 1879. This has allowed by to definitively put together history from the sources of events. The media today has turned to propaganda. So the reliability of the press today on so many things has been lost.

I have learned over the years to go to the contemporary sources and do not rely upon interpretations. The object of life is to learn – not to promote propaganda. I have an extensive collection of newspapers and contemporary accounts from specific events even into ancient times. When I die, I might have enough stuff to really warrant a library compared to pretend presidents pushing their legacy and agenda. I have made a fortune over the years, but I have plowed it back into research since the 1970s.