From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: coax (3174) throughput Newsgroups: bit.listserv.ibm-main Date: 6 Apr 2011 05:47:55 -0700R.Skorupka@BREMULTIBANK.COM.PL (R.S.) writes:
I never measured 3174 ... the 3274 had the opposite problem ... not only did moving lot of electronics out of the head back to shared control unit ... enormously increase coax cable chatter and slow down thruput (i.e. both the amount of chatter on the coax as well as latency for all the back and forth to support really "dumbed down" 3278) ... but the slow electronics in the 3274 had significant hit on (bus&tag) channel busy (transfer rate 640kbytes on the channel side ... but really slow handsaking made raw transfer rate only small part of the channel busy ... analogous to all the really slow handshaking on the coax side enormously slowing down effective response time and transfer rate).
I had done a project for the IMS group when STL was bursting at the
seams and 300 were being put at remote site ... with datacenter support
back to STL. They had tested "remote" 3278 support back to STL and found
it truely horrible and totally unacceptable ... local channel attach
3278 were bad enough having hard time making subsecond response
https://www.garlic.com/~lynn/2001m.html#19 3270 protocol
but for "remote" 3278, it wasn't even "remotely" possile :-)
The side effect of doing support for channel-extender ... allowing "channel" attached 3274s controllers to put at the remote location (and providing 3278 response at remote location, that was indistinguishable from local channel attach), the channel-extender boxes had significantly faster channel interface processing ... getting the 3274s off the real channels improved local processor thruput by 10-15%.
When 3278s originally came out ... we complained loudly to the product group about 3278 interactive performance vis-a-vis 3277. Eventually the product group came back with the reply that 3278s weren't designed for interactive computing but for "data entry" (aka basically online "upgrade" for card punch machines).
The controller channel busy overhead (independent of raw transfer rate)
was to raise its head again with 3090 and 3880 disk controllers (3mbyte
transfer rate). The 3880 channel busy overhead turned out to be so high,
3090 product realized that it had to add a whole bunch additional
channels ... which resulted in having to add an extra TCM to 3090
manufacturing (there were jokes that the 3090 group was going to bill
the 3880 product group for the cost of the increased 3090 manufacturing
cost). This was sort of the leading edge of theme that mainframes with
enormous number of channels being a good thing (when it was actually to
compensate for the channel/controller interface design and slow
controllers would drastically reduce channel effectiveness). a couple
recent posts:
https://www.garlic.com/~lynn/2011.html#37 CKD DASD
https://www.garlic.com/~lynn/2011e.html#15 At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: Itanium at ISSCC Newsgroups: comp.arch Date: Wed, 06 Apr 2011 09:23:58 -0400Anne & Lynn Wheeler <lynn@garlic.com> writes:
the 90s failures involved parallelization library software with APIs where application programmers had to design parallel applications and used the APIs to distributed the work. the issue wasn't just the overhead involved in the libraries distributing the parallelized work ... but the skill level required for programmers doing the (re-engineered) applications.
the "newer" generation of technology from a couple years ago ... involved taking high level business process specification and generating fine-grain SQL statements. The enormous amount of effort that has gone into RDBMS parallelization was then leveraged to gain high-throughput. The automated translating high-level business processs specification to SQL also significantly reduced the application programmer skill level.
so the new ten-core/20-thread XEONs from yesterday (allowing up to 256 sockets per configuration) ... could make one gangbuster RDBMS machine (especially running mission critical applications running enormous numbers of fine-grain SQL statements)
Intel Xeon E7 processor formula for mission-critical computing
http://www.physorg.com/news/2011-04-intel-xeon-e7-processor-formula.html
Intel Unveils 10-Core Xeons, Mission-Critical Servers - HotHardware
http://hothardware.com/Reviews/Intel-Unveils-10Core-Xeons-MissionCritical-Servers/
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 06 Apr, 2011 Subject: Car models and corporate culture: It's all lies Blog: Greater IBMI got roped into going to some of the auto industry C4 taskforce meetings in the early 90s about how to completely remake themselves. They talked about how the standard industry process took 7-8 yrs to come out with new car model ... with minor cosmetic changes for new model years (between "real" new models). Sometime they would have parallel new models with 3-4 yr offset. The issue was that foreign competition had earlier dropped to 3-4yrs to do a new model, was at 18months ... and appeared to be in the process of dropping to less than traditional annual model year. There were also some of the "mainframe" brethren at the meetings & I would chide them offline about what were they going to contribute to the effort ... since their product cycle (at the time) was nearly the same as the US auto industry.
In any case, the C4 meetings highlighted that the foreign competition was able to re-act to changing consumer preferences and leverage newer technology possibly ten times faster than the US auto industry. Also, even though they understood what the problem was and could articulate everything that needed to be done ... nearly two decades later they hadn't made significant changes (enormous vested interests holding back change).
related from IBM Jargon:
Mongolian Hordes Technique - n. A software development method
whereby large numbers of inexperienced programmers are thrown at a
mammoth software project (instead of deploying a small team of skilled
programmers). First recorded in 1965, but popular as ever in the
1990s.
... snip ...
I had sponsored Boyd's briefings at IBM ... which included significant
amount about lean & agile ... recent posts in similar (linkedin) Boyd
group discussion:
https://www.garlic.com/~lynn/2011e.html#90
https://www.garlic.com/~lynn/2011e.html#92
IBM Jargon use to be purely internal document ... but has leaked out on to the internet and search engine finds numerous copies. Mongolian hordes technique was a pejorative reference to the practice.
In boyd's briefings, he would characterize the practice as epidemic in US business and a result of US entry into WW2. The scenario was that the US had to deploy huge inexperienced numbers and to leverage the few skilled resources available ... created a rigid, top-down, command&control structure ... the US WW2 strategy to win was the rigid, top-down, command&control structure using enormous overwhelming resources. Later the methodology was duplicated as the former young army officers moved up the corporate business ladder. Even when there were qualified, experienced people available ... they would be augmented with large number of unqualified and everybody treated as inexperienced and needing rigid, top-down, command&control structure.
misc. past references to Boyd (including being credited with battle
strategy for the conflict in the early 90s ... as well as comments
that the problem with the current round of conflicts was that Boyd had
died in 1997)
https://www.garlic.com/~lynn/subboyd.html
Note that the change in IBM may have been accelerated by the Future
System disaster ... this reference
https://people.computing.clemson.edu/~mark/fs.html
has quote from Chapter 3 of Charles Ferguson and Charles Morris,
Computer Wars: The Post-IBM World, Times Books, 1993. An excerpt:
Most corrosive of all, the old IBM candor died with F/S. Top
management, particularly Opel, reacted defensively as F/S headed
toward a debacle. The IBM culture that Watson had built was a harsh
one, but it encouraged dissent and open controversy. But because of
the heavy investment of face by the top management, F/S took years to
kill, although its wrongheadedness was obvious from the very
outset. "For the first time, during F/S, outspoken criticism became
politically dangerous," recalls a former top executive.
... snip ...
another quote:
...perhaps the most damaging, the old culture under Watsons of free
and vigorous debate was replaced with sycophancy and make no waves
under Opel and Akers.
... snip ...
and the FS failure cast a dark shadow over the corporation for
decades. misc. past posts
https://www.garlic.com/~lynn/submain.html#futuresys
during "FS", it may not have been particularly career enhancing to ridicule "FS" ... including drawing comparisons with a cult film that had been playing continuously in central sq.
another "FS" reference here:
http://www.jfsowa.com/computer/memo125.htm
there is also some discussion of FS here:
https://www.ecole.org/en/session/49-the-rise-and-fall-of-ibm
https://www.ecole.org/en/session/49-the-rise-and-fall-of-ibm
from IBM Jargon:
FS - n. Future System. A synonym for dreams that didn't come
true. That project will be another FS. Note that FS is also the
abbreviation for functionally stabilized, and, in Hebrew, means zero,
or nothing. Also known as False Start, etc.
... snip ...
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 06 Apr, 2011 Subject: History of APL -- Software Preservation Group Blog: Greater IBMre:
Originally, SEs got quite a bit of training, effectively with
"apprentice" type of program to more experienced SEs at customer site
(for some customers, 20 or more on site). With the 23Jun69 unbundling
announcement, not only was the start of charging for application
software ... but also for SE services (for SEs times spent at
customer), putting an end to the SE "apprentice" program. misc. past
posts mentioning 23jun69 unbundling
https://www.garlic.com/~lynn/submain.html#unbundle
As substitute, several cp67 virtual machine datacenters were created to give (branch office) SEs "hands-on" practice with operating systems (running in virtual machines). In addition to CP67, virtual machines, CMS, GML, the internal network and a bunch of other stuff, the science center also ported APL\360 to CMS for CMS\APL ... fixing a lot of problems with allowing large demand-paged virtual memory workspaces (significantly larger than the traditional 16kbytes to 32kbytes that were common with apl\360) ...and providing API for accessing CMS services. This enabled a whole new generation of "real-world" APL applications.
Part of this was growing number of HONE APL-based sales&marketing
support applications ... which eventually came to dominate all HONE
activity (and the original HONE purpose died off) ... later HONE moved
from cp67/cms to vm370/cms & APL\CMS. misc. past posts mentioning HONE
and/or APL
https://www.garlic.com/~lynn/subtopic.html#hone
The science center also made its cp67/cms service available to other internal locations as well as students and staff at educational institutions in the cambridge area. A major, early CMS\APL application was from Armonk when the business planning people loaded the most valuable corporate asset on the science center cp67 system ... and developed customer business modeling and planning applications in APL (this also tested the cp67 security paradigm keeping all the univ. students away from the most valuable of corporate information assets).
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: Cool Things You Can Do in z/OS Newsgroups: bit.listserv.ibm-main Date: 6 Apr 2011 12:22:18 -0700sam@PSCSI.NET (Sam Siegel) writes:
when we were doing HA/CMP product for the company ... I had coined the
terms disaster survivability and geographic
survivability (to differentiate from disaster/recovery). I had
also been asked to write a section for the corporate continuous
availability strategy document ... but it got pulled when both
Rochester (as/400) and POK (mainframe) complained (that they couldn't
meet the requirements). We also did some work with the 1-800 service
(database service that maps 1-800 numbers to the "real" exchange
number ... required five-nines availability). misc. past posts
mentioning ha/cmp
https://www.garlic.com/~lynn/subtopic.html#hacmp
later we had some dealings with one of the large financial transaction infrastructures ... and they attributed their multiple year, 100% availability to
• geographically separated, replicated IMS hot-standby operation • automated operator
i.e. as hardware has became much more reliable ... unscheduled outages came to be dominated by environmental issues/outages and human mistakes
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: Cool Things You Can Do in z/OS Newsgroups: bit.listserv.ibm-main Date: 6 Apr 2011 13:07:37 -0700tony@HARMINC.NET (Tony Harminc) writes:
there was an incident a few years back where the 1-800 mapping for major percentage of the POS (point-of-sale) card-swipe terminals in the US was down for 12 minutes during a mid-day period ... this was treated as a serious corporate incident between major transaction processor and major telco operation.
five-nines availability is something like 5min (total) outage per year (includes both scheduled and unscheduled).
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 08 Apr, 2011 Subject: New job for mainframes: Cloud platform Blog: MainframeZoneNew job for mainframes: Cloud platform
from ablve:
As companies take steps to develop private clouds, mainframes are
looking more and more like good places to house consolidated and
virtualized servers. Their biggest drawback? User provisioning is
weak.
... snip ...
I've repeatedly mentioned that virtual machine based cloud operations
go back to the 60s ... some past posts
https://www.garlic.com/~lynn/submain.html#timeshare
... and the largest such operation in the 70s & 80s was internal,
world-wide sales & marketing support HONE system. misc. past posts
mentioining HONE
https://www.garlic.com/~lynn/subtopic.html#hone
CP67 did a couple things in the 60s to help open up 7x24 operation.
At the time, mainframes were leased and had monthly shift charges based on the processor meter (which ran whenever the processor &/or channels were active). Early deployments tended to have relatively light off-shift usage. One of the tricks was a terminal channel program sequence that left the line open to accept incoming characters but wouldn't run the channel (& processor meter) when no characters were arriving.
Another was significantly improving operator-less/dark-room off-shift operation to minimize operation costs during light off-shift operation.
CP67 was enhanced to automatically take a "dump" (to disk) and
re-ipl/re-boot after a failure ... coming back up and available for
service. One of the issues was that the growing number of service
virtual machines (virtual appliances) still required manual
restart. I then did the "autolog" command, originally for automatic
benchmarking (could run large number of unattended benchmarks with
system reboot between operation) ... discussed here:
https://www.garlic.com/~lynn/2010o.html#48
It then started being used for automatic startup of service virtual
machines ... and after conversion from cp67 to vm370 ... the product
group then picked up a number of CSC/VM features for VM370 release
3. old email refs:
https://www.garlic.com/~lynn/2006w.html#email750102 ,
https://www.garlic.com/~lynn/2006w.html#email750827
This (recent) post in (linkedin) IBM Historic Computing group
discusses support for loosely-coupled (cluster), single-system-image,
load-balancing & fail-over support done during the 70s by a number of
large virtual-machine-based service operation (including HONE). Also,
in the 70s, at least one virtual-machine based online commercial
service bureau provided for migrating active users between processors
in loosely-coupled (cluster) configuration ... supported
non-disruptive removal of processor in cluster for things like
scheduled downtown for preventive maintenance.
https://www.garlic.com/~lynn/2011e.html#79
In the mid-70s, the internal US HONE datacenters had been consolidated in silicon valley. Then in the early 80s, somewhat in response to earthquake ... the HONE cluster support was extended with a replicated datacenter in Dallas and then a 3rd in Boulder.
another side of the coin:
Cloud Use Rises, Mainframe Usage Declines as Data Centers Grow and
Green, According to AFCOM Survey EON: Enhanced Online News
http://eon.businesswire.com/news/eon/20110330005393/en/cloud/disaster-recovery/data-center
some more in this server market segment comparison between 2002 & 2010
(part of yesterday's xeon announce):
Intel Unveils 10-Core Xeons, Mission-Critical Servers - HotHardware
http://hothardware.com/Reviews/Intel-Unveils-10Core-Xeons-MissionCritical-Servers/
other details from that announce:
Performance + Reliability + Security = Intel Xeon Processor Formula
for Mission-Critical Computing
http://newsroom.intel.com/community/intel_newsroom/blog/2011/04/05/performance-reliability-security-intel-xeon-processor-formula-for-mission-critical-computing?cid=rss-258152-c1-265964
IBM Jumps Into Cloud, Customers Tip-toe Behind
http://news.yahoo.com/s/pcworld/20110408/tc_pcworld/ibmjumpsintocloudcustomerstiptoebehind
a related recent discussion in (linkedin) Greater IBM discussion "History of APL -- Software Preservation Group" ... regarding an important part of IBM SE training was "apprentice" type program for new "SEs" in group of more experienced SEs onsite at customer accounts. This was cutoff with the 23Jun69 unbundling announcement that started to charge for SE services (as well as application software); requirement that individual SE time at customer accounts had to be charged for ... and couldn't figure out charging policy for the "apprentice" SEs.
This was the original motivation for internal HONE system (the largest "cloud" service in the 70s & 80s) ... several cp67 (virtual machine) datacenters with online, remote access for SEs in branch offices ... so that they could practice their operating system skills.
part of that recent thread:
https://www.garlic.com/~lynn/2011f.html#3
misc. past posts mentioning unbundling
https://www.garlic.com/~lynn/submain.html#unbundle
another cloud related item:
Facebook Opens Up Its Hardware Secrets; The social network breaks an
unwritten rule by giving away plans to its new data center--an action
it hopes will make the Web more efficient
http://www.technologyreview.com/news/423570/facebook-opens-up-its-hardware-secrets/?p1=MstRcnt&a=f
for "HONE" related trivia ... upthread mentions that in mid-70s, the US HONE datacenters were consolidated in silicon valley. Do online satellite map search for Facebook's silicon valley address ... the bldg. next to it was the HONE datacenter (although the bldg. has a different occupant now).
Note that environmental and people mistakes had come to dominate outages quite some time ago. Countermeasures were clusters (loosely-coupled) as well geographically separated operation. Applications that provide for fall-over/take-over to handle local outages ... also can be used to mask increasingly rare local hardware faults.
Long ago and far away, my wife had been con'ed into going to POK to be
in charge of (mainframe) loosely-coupled (cluster) architecture. While
there she did Peer-Coupled Shared Data architecture ... but except for
IMS hot-standby, there was little uptake (until sysplex & parallel
sysplex) ... which contributed to her not remaining long in the
position (that and battles with the communication group trying to
force her into using SNA for loosely-coupled operation). misc. past
posts
https://www.garlic.com/~lynn/submain.html#shareddata
A few years ago we were dealing with one of the largest financial transaction operation and they attributed their several year 100% availability to
• IMS hot-standby (at geographically separated sites) • automated operator
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: New job for mainframes: Cloud platform Newsgroups: bit.listserv.ibm-main Date: 8 Apr 2011 06:02:51 -0700New job for mainframes: Cloud platform
from above:
As companies take steps to develop private clouds, mainframes are
looking more and more like good places to house consolidated and
virtualized servers. Their biggest drawback? User provisioning is weak.
... snip ...
also:
http://lnkd.in/F6X_3Y
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 08 Apr, 2011 Subject: New job for mainframes: Cloud platform Blog: MainframeZonere:
also:
http://lnkd.in/F6X_3Y
post from two years ago: "From The Annals of Release No Software
Before Its Time"
https://www.garlic.com/~lynn/2009p.html#43
referring to an announcement regarding z/VM clustering (HONE had done
vm clustering including geographic distributed, 30yrs earlier):
http://www.vm.ibm.com/zvm610/zvm61sum.html
and also purescale announcement:
http://www-03.ibm.com/press/us/en/pressrelease/28593.wss
referring to this regarding RDBMS scale-up in Jan92
https://www.garlic.com/~lynn/95.html#13
when the mainframe DB2 group commented that if I was allowed to go ahead, it would be at least five yrs ahead of them. Within a couple weeks, the effort was transferred, announced as supercomputer for numerical intensive, technical & scientific *only*, and we were told we couldn't work on anything with more than four processors.
It was about the same time I had been asked to write a section for the corporate continuous availability strategy document ... but it got pulled when both rochester (as/400) and POK (mainframe) complained (that they couldn't meet the objectives).
Some more in (linkedin) Greater IBM discussion about Watson Ancestors
... archived here:
https://www.garlic.com/~lynn/2011d.html#7 ,
https://www.garlic.com/~lynn/2011d.html#24 ,
https://www.garlic.com/~lynn/2011d.html#29 ,
https://www.garlic.com/~lynn/2011d.html#40
Mainframe SQL/DS came direct from System/R (original SQL/RDBMS
implementation) ... and mainframe DB2 is descendent of both SQL/DS and
System/R ... misc. past posts mentioning System/R
https://www.garlic.com/~lynn/submain.html#systemr
At the time we were doing HA/CMP ... non-mainframe DB2 (code-name shelby) was being written in C for OS/2, so we were working with major non-mainframe RDBMS; Ingres, Sybase, Informix, and Oracle. Most had some form of VAX/cluster implementations. In working with them ... providing a vax/cluster-like API for the distributed lock manager (DLM) seemed to the fastest way to shipping cluster scale-up products. Going about the DLM implementation, the various RDBMS vendors had strong feelings about things that could be significantly improved in the VAX/cluster implementation (for throughput, scale-up, and recovery) ... in addition to the experience I already had doing a number of cluster efforts.
Part of the mainframe issue was that the massive new (mainframe) DBMS effort in STL had been EAGLE ... which pretty much allowed System/R and SQL/DS activity to go unnoticed. It wasn't until EAGLE effort was crashing ... that they asked how fast could a (mainframe/MVS) RDBMS be turned out ... resulting in MVS/DB2 not shipping until 1983 (making it fairly late to the RDBMS game)
some recent posts discussing EAGLE, System/R, SQL/DS, RDBMS
https://www.garlic.com/~lynn/2011d.html#42
https://www.garlic.com/~lynn/2011d.html#52
https://www.garlic.com/~lynn/2011e.html#16
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: New job for mainframes: Cloud platform Newsgroups: bit.listserv.ibm-main Date: 8 Apr 2011 09:02:54 -0700ibm-main@TPG.COM.AU (Shane Ginnane) writes:
previous was from several days ago ... more recent items from today:
IBM Jumps Into Cloud, Customers Tip-toe Behind
http://www.pcworld.com/businesscenter/article/224688/ibm_jumps_into_cloud_customers_tiptoe_behind.htm
IBM Forecasts $7 Billion In Cloud Revenue
http://www.informationweek.com/news/global-cio/interviews/showArticle.jhtml?articleID=223800165
there is also some additional discussion in the linkedin mainframe group
URL
http://lnkd.in/F6X_3Y
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 09 Apr, 2011 Subject: History of APL -- Software Preservation Group Blog: Greater IBMre:
quick search turns up this reference as follow-on (mentions STAIRS
source code was being made available to ease customer conversion):
http://www-01.ibm.com/software/data/sm370/about.html
web search turns up a number of products/services that had been based on STAIRS.
code fixes/PTFs was (field engineering) "RETAIN" ... also available from the branch office terminals.
This is old post with PARASITE/STORY for automatic (remote) logon to
RETAIN ... and fetch. PARASITE/STORY was an internal application that
ran on CMS (utilizing VM370's virtual device/terminal support) and
provided HLLAPI-like scripting capability (before IBM/PCs)
https://www.garlic.com/~lynn/2001k.html#36
Old post mentioning somebody printed an "April 1st" memo (about
passwords) on corporate letter head paper on one of the 6670s (started
out as ibm copier/3 with computer interface) in bldg. 28 and put them
in all the bldg. bulletin boards.
https://www.garlic.com/~lynn/2001d.html#53
bldg. 28 also had project to enhance 6670 to support APA6670/sherpa
(all-points-addressable ... be able to do images, also supporting 6670
as scanner)
https://www.garlic.com/~lynn/2006p.html#email820304
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 09 Apr, 2011 Subject: History of APL -- Software Preservation Group Blog: Greater IBMre:
recent post (in linkedin Boyd group):
https://www.garlic.com/~lynn/2011e.html#92
also
http://lnkd.in/j9U4bS
with PROFS entry from IBM Jargon:
PROFS - profs n. Professional Office System. A menu-based system that
provides support for office personnel such as White House staff, using
IBM mainframes. Acclaimed for its diary mechanisms, and accepted as
one way to introduce computers to those who don't know any better. Not
acclaimed for its flexibility. PROFS featured in the international
news in 1987, and revealed a subtle class distinction within the ranks
of the Republican Administration in the USA. It seems that Hall, the
secretary interviewed at length during the Iran-Contra hearings,
called certain shredded documents PROFS notes as do IBMers who use the
system. However, North, MacFarlane, and other professional staff used
the term PROF notes. v. To send a piece of electronic mail, using
PROFS. PROFS me a one-liner on that. A PROFS one-liner has up to one
line of content, and from seven to seventeen lines of boiler
plate. VNET
... snip ...
the PROFS entry is also quoted in this post about SNA/VTAM
misinformation
https://www.garlic.com/~lynn/2011e.html#57
in the late 80s when the communication group was lobbying to convert
the internal network to SNA/VTAM one of the things they did was tell
the top executives that PROFS was a VTAM application. Above thread
also references some old email of theirs claiming that SNA/VTAM was
also applicable for NSFNET backbone (i.e. operational precursor to the
modern internet, aka tcp/ip). misc. past posts:
https://www.garlic.com/~lynn/subnetwork.html#nsfnet
The author of PARASITE/STORY (upthread reference) also was the author of VMSG. The PROFS group used a very early copy of VMSG for their email client. When the VMSG author offered the PROFS group a current/enhanced copy of VMSG, they tried to get him fired (having claimed to have done all of PROFS themselves). Things quieted down after the VMSG author pointed out that every PROFS message in the world carried his initials in a non-displayed field. After that, he only shared the VMSG source with two other people.
a little x-over from the "Car Models and corporate culture: It's all
lies" ... also archived here
https://www.garlic.com/~lynn/2011f.html#2
one could make the case that the aftermath of FS with no dissension was still active (at least) throughout the 80s and 90s.
In the mid-80s, top executives were predicting that revenue was going to double from $60B to $120B ... and there was massive program to build-out manufacturing capacity (mostly mainframe related, to meed the projected demand). However, at the time, it was relatively trivial to show that things were heading in the opposite direction (and the company going into the red a few years later)
and as mentioned upthread, the disk division comment about communication group being responsible for the demise of the disk division.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: New job for mainframes: Cloud platform Newsgroups: bit.listserv.ibm-main Date: 11 Apr 2011 07:38:12 -0700timothy.sipples@US.IBM.COM (Timothy Sipples) writes:
when I was undergaduate in the 60s, some people from the science center came out and installed (virtual machine) cp67 on the 360/67 (as alternative to tss/360). cp67 had "automatic" terminal identification for 1052 & 2741 ... playing games switching the line-scanners with the 2702 SAD command. The univ. had bunch of TTY/ascii terminals ... so I set out to add TTY/ascii support also doing automatic terminal identification. It almost worked ... being able to dynamically identify 1052, 274, & TTY for directly/fixed connect lines.
I had wanted to have a single dial-up number for all termainls ... with "hunt-group" ... allowing any terminal to come in on any port. The problem was that the 2702 took a short-cut and hardwired the line-speed for each port. This somewhat prompted the univ. to do a clone controller effort ... to dynamically do both automatic termeinal & automatic speed determination (reverse engineer channel interface, build controller interface board and program minicomputer to emulate 2702).
Two early "bugs" that stick in my mind ...
1) the 360/67 had high-speed location 80 timer ... and if the channel interface board held the memory bus for two consecutive timer-tics (a timer-tic to update location 80 was stalled because memory bus was held ... and the next timer-tic happened while the previous timer-tic was still pending), the processor would stop & redlight
2) initial data into memory was all garbage. turns out had overlooked bit memory order. minicomputer convention was leading (byte) bit off the line started off into high-order (byte) bit position ... while 2702 line-scanner convention was to place leading (byte) bit off the line in the lower order (byte) bit position. while the minicomputer then was placing data into memory in line-order bit positiion ... each byte had the bit order reversed compared to the 2702 convention (standard 360 ascii translate tables that I had borrowed from BTAM handled the 2702 bit-reversed bytes).
... later, four of us get written up for being responsible for some portion of the mainframe clone controller business. A few years ago, in large datacenter, I ran across a descendent of our original box, handling a major portion of the dial-up POS cardswipe terminals in the country (some claim that it still used the original channel interface board design).
I had posted same cloud item in a number of linkedin mainframe group
https://www.garlic.com/~lynn/2011f.html#6 New job for mainframes: Cloud platform
https://www.garlic.com/~lynn/2011f.html#8 New job for mainframes: Cloud platform
also
http://lnkd.in/F6X_3Y
also refers to internal (virtual machine) HONE system being the largest
"cloud" operation in the 70s & 80s. In the mid-70s, the US HONE
datacenters were consolidated in silicon valley ... where it created the
largest single-system-image cluster operation. Then in the early 80s,
because of earthquake concerns, it was replicated in Dallas ... with
distributed, load-balancing and fall-over between Dallas & PaloAlto
... eventually growing to 28 3081s. misc. past posts mentioning HONE
https://www.garlic.com/~lynn/subtopic.html#hone
HONE also discussed in this linkedin Greater IBM (current & former IBM
employee) group about APL software preservation (major portion of HONE
applications supporting worldwide sales & marketing had been implemented
in APL; numerous HONE-clones all around the world):
https://www.garlic.com/~lynn/2011e.html#83 History of APL -- Software Preservation Group
https://www.garlic.com/~lynn/2011f.html#3 History of APL -- Software Preservation Group
https://www.garlic.com/~lynn/2011f.html#10 History of APL -- Software Preservation Group
https://www.garlic.com/~lynn/2011f.html#11 History of APL -- Software Preservation Group
another cloud related item:
Facebook Opens Up Its Hardware Secrets; The social network breaks an
unwritten rule by giving away plans to its new data center--an action it
hopes will make the Web more efficient
http://www.technologyreview.com/news/423570/facebook-opens-up-its-hardware-secrets/?p1=MstRcnt&a=f
for "HONE" related trivia ... the silicon valley HONE datacenter; do online satellite map search for Facebook's silicon valley address ... the bldg. next to it was the HONE datacenter (although the bldg. has a different occupant now).
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 11 Apr, 2011 Subject: Car models and corporate culture: It's all lies Blog: Greater IBMre:
a little x-over from "History of APL" discussion about mid-80s when
top executives were predicting that world-wide revenue would double
from $60B to $120B .... mostly mainframe ... even when it was fairly
trivial to show business was heading in opposite direction (and
company goes into the red just a few yrs later). Part of the strategy
was a massive manufacturing building program ... apparently to double
the manufacturing capacity (again mostly mainframe related).
https://www.garlic.com/~lynn/2011e.html#83 History of APL -- Software Preservation Group
https://www.garlic.com/~lynn/2011f.html#3 History of APL -- Software Preservation Group
https://www.garlic.com/~lynn/2011f.html#10 History of APL -- Software Preservation Group
https://www.garlic.com/~lynn/2011f.html#11 History of APL -- Software Preservation Group
However, past discussions that "fast track" program also became
especially epidemic during the period (possibly trying to also double
the number of executives) ... from IBM Jargon:
fast track - n. A career path for selected men and women who appear to
conform to the management ideal. The career path is designed to
enhance their abilities and loyalty, traditionally by rapid promotion
and by protecting them from the more disastrous errors that they might
commit.
... snip ...
which seemed to have side-effect that large number of executives spending massive amount of their time & effort "managing" their careers (as opposed to day-to-day corporate & business activity).
other recent posts mentioning "fast track":
https://www.garlic.com/~lynn/2011.html#55 America's Defense Meltdown
https://www.garlic.com/~lynn/2011d.html#1 If IBM Hadn't Bet the Company
https://www.garlic.com/~lynn/2011d.html#3 If IBM Hadn't Bet the Company
https://www.garlic.com/~lynn/2011d.html#6 If IBM Hadn't Bet the Company
https://www.garlic.com/~lynn/2011d.html#12 I actually miss working at IBM
https://www.garlic.com/~lynn/2011d.html#24 IBM Watson's Ancestors: A Look at Supercomputers of the Past
https://www.garlic.com/~lynn/2011d.html#78 I actually miss working at IBM
https://www.garlic.com/~lynn/2011d.html#79 Mainframe technology in 2011 and beyond; who is going to run these Mainframes?
https://www.garlic.com/~lynn/2011e.html#45 junking CKD; was "Social Security Confronts IT Obsolescence"
https://www.garlic.com/~lynn/2011e.html#90 PDCA vs. OODA
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 11 Apr, 2011 Subject: How is SSL hopelessly broken? Let us count the ways Blog: FacebookHow is SSL hopelessly broken? Let us count the ways
we had been called in to consult with small client/server startup that wanted to do payment transactions on their server, they had also invented this technology called SSL they wanted to us (the result is now frequently called "electronic commerce"). By the time we were finished with the deployments ... most of the (current) issues were very evident.
very early in the process I had coined the term "comfort certificate"
... since the digital certificate actually created more problems than
it solved ... in fact, in many cases, it was totally redundant and
superfluous ... and existed somewhat as magic "pixie dust" ... lots of
old posts mentioning SSL digital certificates:
https://www.garlic.com/~lynn/subpubkey.html#sslcert
There was two parts of SSL deployment for "electronic commerce"
... between the browser and webserver and between the webserver and
the "payment gateway" ... I had absolute authority over interface
deployment involving "payment gateway" ... but had only advisery over
browser/webserver. There were a number of fundamental assumptions
related to SSL for browser/webserver secure deployment ... that were
almost immediately violated by merchant webservers (in large part
because of the high overhead of SSL cut their throughput by 85-90%). I
had mandated mutual authentication for webserver/gateway
(implementation didn't exist originally) and by the time deployment
was done the use of SSL digital certificates was purely a side-effect
of the crypto library being used
https://www.garlic.com/~lynn/subnetwork.html#gateway
primary use of SSL in the world today is electronic commerce for hiding payment transaction information. The underlying problem is the transaction information is dual-use both authentication and needed by dozens of business processes at millions of of places around the world.
In the X9A10 financial working group we directly addressed the
dual-use problem with the x9.59 financial standard (directly
addressing the problem, x9.59 is significantly lighter weight than SSL
as well as KISS). This eliminated the need to hide the transaction
details ... also eliminates the threat from majority of data breaches
(doesn't eliminate data breaches, just eliminates crooks being able to
use the information for fraudulent purposes). Problem (as always) is
there is significant vested interests in the current status quo.
https://www.garlic.com/~lynn/subpubkey.html#x959
we were also tangentially involved in the cal. state data breach
notification legislation. we had been brought in to help wordsmith the
cal. state electronic signature legislation. many of the parties were
involved in privacy issues and had done detailed, indepth public
surveys and found #1 issue was identify theft, and major form was
account fraud as a result of data breaches. there was little or
nothing being done about such breaches (security measures are
nominally used to protect party involved, breaches weren't affecting
those with the data). It was hoped that publicity from breach
notification might motivate the industry to take countermeasures
https://www.garlic.com/~lynn/subpubkey.html#signature
crooks used consumer transaction information from breach to perform fraudulent transactions against consumer accounts. the merchants and financial institutions that had the breach didn't have any (direct) problems (until the breach notification law made it public).
x9.59 financial transaction standard addressed it in totally different way ... it eliminated crooks being able to use information to perform fraudulent transactions ... so it eliminated needing to hide the information ... so it eliminated the need to use SSL to hide the information ... and it eliminated the financial motivation for crooks to perform breaches (since the information was of no practical use).
aka breach notification came about because enormous amount of fraud was happening to consumers as a result of the breaches ... and most of those that needed to protect/hide the data couldn't care less (they were doing little or nothing to prevent breaches). x9.59 financial standard addressed it differently, it eliminated the ability for crooks to use the information to perform fraudulent transactions.
With regard to x9.59 standard, all software would support whatever is required (like starbucks accepting barcode generated by iphone app). X9.59 specification was originally done by financial industry to address enormous fraud. However, large institutions in US turned out to be getting approx. half their bottom line from "payment transaction fees" (charged merchants) that had been set proportional to fraud. Eliminating that fraud could cut the related fees by better than order of magnitude ... also basically commoditizes payment industry and lowers barrier to entry for competition. Eliminating nearly all such fraud (and drastically reducing fees) met enormous resistance from bank business people (effectively making significant profit from fraud). Mistake is thinking of x9.59 as a technology solution ... it is a business specification ... doesn't directly involve consumers ... there are bookshelves of current specifications that consumers aren't aware of.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: Identifying Latest zOS Fixes Newsgroups: bit.listserv.ibm-main Date: 11 Apr 2011 15:33:11 -0700mike.a.schwab@GMAIL.COM (Mike Schwab) writes:
much of desktop evolved in purely stand-alone enviornment ... with some early (3270) terminal emulation. then was added, private, business, closed, "safe", network support ... with lots of applications & files that included automated "scripted" enhancements. At the 1996 MDC (held at Moscone) there were huge number of banners about moving to internet (simple remapping of the networking conventions w/o the corresponding countermeasures involved moving from "safe" environment to extremely "hostile" environment; periodic analogy with going out airlock into open space w/o space suit).
However, the constant subtheme (at '96 MDC) was "protecting your investment" ... referring to all the scripting capability. Starting early part of this century, such exploits began to eclipse the buffer length problems ... along with the heavy weight security convention of analyze/filtering incoming files against an enormously bloated library of possible exploit "signatures"
old post doing word frequency analysis of CVE "bug" reports ... and
suggesting to mitre that they require a little more formal structure
in the reports (at the time, I got pushback that they were lucky to
get any reasonable verbage):
https://www.garlic.com/~lynn/2004e.html#43 security taxonomy and CVE
more recent reference to CVE ... which has since moved to NIST
https://www.garlic.com/~lynn/2011d.html#8 Security flaws in software development
note that the original mainframe tcp/ip protocol stack had been done
in vs/pascal ... and suffered none of the buffer length exploits found
in C-language implementations. there were other thruput and pathlength
issues with that implementation ... but I did the RFC1044 enhancements
for the implementation ... and in some testing at Cray Research
... got sustained channel media throughput between Cray and 4341
... using only modest amount of 4341 processor. misc. past posts
mentining RFC1044 support
https://www.garlic.com/~lynn/subnetwork.html#1044
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: Jean Bartik, "Software" Pioneer, RIP Newsgroups: alt.folklore.computers Date: Mon, 11 Apr 2011 23:44:38 -0400Louis Krupp <lkrupp_nospam@indra.com.invalid> writes:
recent posts mentioning the above:
https://www.garlic.com/~lynn/2011d.html#1 If IBM Hadn't Bet the Company
https://www.garlic.com/~lynn/2011d.html#12 I actually miss working at IBM
https://www.garlic.com/~lynn/2011e.html#90 PDCA vs. OODA
https://www.garlic.com/~lynn/2011e.html#92 PDCA vs. OODA
recent posts (also) referencing the culture:
https://www.garlic.com/~lynn/2011f.html#2 Car models and corporate culture: It's all lies
https://www.garlic.com/~lynn/2011f.html#13 Car models and corporate culture: It's all lies
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: New job for mainframes: Cloud platform Newsgroups: bit.listserv.ibm-main Date: 12 Apr 2011 07:48:47 -0700graeme@ASE.COM.AU (Graeme Gibson) writes:
yes, well ... I thot it was also interesting that 2702 (IBM line scanners) managed to (also) reverse bits within bytes (before x86 even appeared on the scene).
other trivia was HP had major hand in Itanium (designed to be dual-mode ... both big-endian & little-endian) ... which at one time was going to be the "mainframe killer" ... since then lots of Itanium business critical features have been migrated to XEON chips (and various recent news items projecting death of Itanium).
major person behind wide-word & Itanium had earlier been responsible
for 3033 dual-address space mode ... retrofitted a little of 370-xa
access registers to 3033 to try and slow the exploding common segment
problem (with 24-bit, 16mbyte virtual address space ... and MVS kernel
image taking half of each virtual address space ... large installations
were approaching situation where CSA was going to be 6mbytes
... reducing space for applications to 2mbytes). itanium stuff
http://www.hpl.hp.com/news/2001/apr-jun/worley.html
other pieces from wayback machine:
https://web.archive.org/web/20010722130800/www.hpl.hp.com/news/2001/apr-jun/2worley.html
https://web.archive.org/web/20000816002838/http://www.hpl.hp.com/features/bill_worley_interview.html
internal IBM had some critical chip-design tools implemented in Fortran running on large number of carefully crafted MVS systems ... and were having increasingly difficult time to keep the application under 7mbytes (MVS kernel image at 8mbytes and minimum CSA size was 1mbyte, leaving maximum of 7mbytes for applications) ... they were being faced with having to convert the whole operation to vm/cms ... since that would allow them to have nearly whole 16mbytes for application.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 12 Apr, 2011 Subject: 21st century India: welcome to the smartest city on the planet Blog: Greater IBM21st century India: welcome to the smartest city on the planet
from above:
From rocket science to DNA research, India is ridding itself of its
poor country image. In this extract from her book Geek Nation, Angela
Saini visits Lavasa, an emerging electronic 'dream city'
... snip ...
In the 90s, there was report that half of the US advanced technology (STEM) degrees were to asia & india ... and it would only take slight changes in relative relation between their home economy and the US economy to reach tipping point and they return to their home country. US had whole technology sectors dominated by US graduates born in asia and india, that had their education paid for by their home governments, were suppose to remain in US industry for 5-7 yrs and were then obligated to return home.
While not so extreme yet, after the recent financial disaster decade ... there are large number of reports about the disappearing US middle class with extreme shift in wealth to the top 1%.
many of the STEM articles ... fail to differentiate the percent of US
graduates that are foreign born and the conditions & percent likely to
return home and/or even obligated to return home
https://en.wikipedia.org/wiki/STEM_fields
that is separate from the large increase in STEM graduates from institutions in Asia.
a little NSF STEM related (I think they are trying to avoid using the
word "crisis")
http://www.nsf.gov/nsb/stem/
http://www.smdeponews.org/funding-opportunities/nsf-seeks-proposals-for-transforming-stem-learning/
http://www.ptec.org/items/detail.cfm?id=9186
http://www.nsf.gov/pubs/2009/nsf09549/nsf09549.pdf
misc. past posts mentioning STEM:
https://www.garlic.com/~lynn/2007s.html#22 America Competes spreads funds out
https://www.garlic.com/~lynn/2010b.html#19 STEM crisis
https://www.garlic.com/~lynn/2010b.html#24 Happy DEC-10 Day
https://www.garlic.com/~lynn/2010b.html#26 Happy DEC-10 Day
https://www.garlic.com/~lynn/2010b.html#56 Happy DEC-10 Day
https://www.garlic.com/~lynn/2010b.html#57 Happy DEC-10 Day
https://www.garlic.com/~lynn/2010b.html#59 Happy DEC-10 Day
https://www.garlic.com/~lynn/2010c.html#87 search engine history, was Happy DEC-10 Day
https://www.garlic.com/~lynn/2010d.html#1 search engine history, was Happy DEC-10 Day
https://www.garlic.com/~lynn/2010f.html#38 F.B.I. Faces New Setback in Computer Overhaul
https://www.garlic.com/~lynn/2010f.html#45 not even sort of about The 2010 Census
https://www.garlic.com/~lynn/2010f.html#84 The 2010 Census
https://www.garlic.com/~lynn/2010g.html#37 16:32 far pointers in OpenWatcom C/C++
https://www.garlic.com/~lynn/2010h.html#47 COBOL - no longer being taught - is a problem
https://www.garlic.com/~lynn/2010i.html#41 Idiotic programming style edicts
https://www.garlic.com/~lynn/2010k.html#18 taking down the machine - z9 series
https://www.garlic.com/~lynn/2010l.html#14 Age
https://www.garlic.com/~lynn/2010m.html#37 A Bright Future for Big Iron?
https://www.garlic.com/~lynn/2010p.html#78 TCM's Moguls documentary series
https://www.garlic.com/~lynn/2010q.html#69 No command, and control
https://www.garlic.com/~lynn/2011.html#80 Chinese and Indian Entrepreneurs Are Eating America's Lunch
https://www.garlic.com/~lynn/2011b.html#0 America's Defense Meltdown
https://www.garlic.com/~lynn/2011c.html#45 If IBM Hadn't Bet the Company
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: Jean Bartik, "Software" Pioneer, RIP Newsgroups: alt.folklore.computers Date: Tue, 12 Apr 2011 13:16:36 -0400re:
past post
https://www.garlic.com/~lynn/2011f.html#44 Happy DEC-10 Day
that in the wake of Jim Gray leaving and his "MIP Envy" tome, there were visits (& trip reports) to a number of other institutions (for comparison)
20Sep80 version of MIP Envy
https://www.garlic.com/~lynn/2007d.html#email800920
in this post
https://www.garlic.com/~lynn/2007d.html#17 Jim Gray Is Missing
also references 24Sep80 version
https://web.archive.org/web/20081115000000*/http://research.microsoft.com/~gray//papers/CritiqueOfIBM%27sCSResearch.doc
this has summary of some of the dataprocessing at some of the visited
(summer 1981) institutions
https://www.garlic.com/~lynn/2001l.html#61 MVS History (all parts)
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: New job for mainframes: Cloud platform Newsgroups: bit.listserv.ibm-main Date: 12 Apr 2011 11:47:12 -0700shmuel+ibm-main@PATRIOT.NET (Shmuel Metz , Seymour J.) writes:
sorry ... access registers were in the "811" architecture documents
(supposedly named for the nov78 date on most of the documents) ... "811"
pieces then leaked out in various machine levels. some of the (3033 &)
3081 discussed in detail here
http://www.jfsowa.com/computer/memo125.htm
... "811", 3033 and 3081 were hurry-up patch up efforts recovering
from the FS disaster ... some more discussion in this (linkedin)
"Greater IBM" (current & former IBMers)
https://www.garlic.com/~lynn/2011f.html#2 Car models and corporate culture: It's all lies
https://www.garlic.com/~lynn/2011f.html#13 Car models and corporate culture: It's all lies
it wasn't until 3090 ... that you start to see a "real" new machine (including vm/cms 4361&3370s being the service processor in all 3090s ... even 3090s that nominally had operating system w/o 3370FBA support).
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 13 Apr, 2011 Subject: WHAT WAS THE PROJECT YOU WERE INVOLVED/PARTICIPATED AT IBM THAT YOU WILL ALWAYS REMEMBER? Blog: Greater IBMGetting blamed for online computer conferencing in the late 70s and early 80s on the internal network ... some of the Tandem Memos were packaged and sent to each member of the executive board (300 pages printed dual-sided on 6670 ... packaged in "Tandem" 3-ring binders ... folklore is that 5of6 immediately wanted to fire me). from IBM Jargon:
even had short article in Nov81 Datamation.
Note that pre-backbone (approaching 1000 nodes) was pretty much
in-spite of the company. There was even an Armonk "investigation" into
the phenomenon ... and it concluded that it couldn't exist. The logic
was that if traditional SNA architecture principles had been used, to
have developed the software for such a network would have required
more resources than was currently available in the whole corporation.
https://www.garlic.com/~lynn/subnetwork.html#internalnet
Of course the internal network used a totally different approach
... and in the late 80s, even layering SNA links into the internal
network still required enormous investment ... both software and
additional hardware ... at that time, it would have been enormously
more cost/effective to have layered it on TCP/IP (rather than SNA);
mention in this (greater ibm) thread:
https://www.garlic.com/~lynn/2011.html#4 Is email dead? What do you think?
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: First 5.25in 1GB drive? Newsgroups: alt.folklore.computers Date: Wed, 13 Apr 2011 16:42:11 -0400from long ago and far away ...
CAPACITY AVG SEEK/TRK-TRK LATENCY RAW DATA RATE (MB) FMT (MS) (MS) (MB/S) HP97540 663 17/3 7.47 2.5 CDC WREN 6 676 16/3 8.33 1.875 CDC RUNNER 330 10.7/4 8.33 1.875 (SHORT STROKE) NEWBURY 9820 690 16/2.6 8.33 1.875 MICROPOLIS 660-890 14/3 8.33 2.5 1590's HITACHI 661 16/4 8.33 2.4 DK515C-78 TOSHIBA 330 18/5 8.33 1.875 MK-250F SIEMENS 660 16/5 8.43 1.875 4400 FUJITSU 660 16/4 8.33 1.875 M2263 FUJITSU 660 16/4 8.33 1.875 **** REDWING 857 12/1.5 6.02 3 MAXTOR 670 18/3 8.33 1.875 (b) 3 1/2in SCSI magnectic disks: Most announcements of 3 1/2in disks are not in the high performance area, like Lighting file. Seagate, Fujitsu, Miniscribe, and Toshiba are in this area (seek time > 20 ms, trk to trk > 6 ms, capacity less than 120 MB). Only two venders announcements have higher performance: CAPACITY AVG SEEK/TRK-TRK LATENCY RAW DATA RATE (MB) FMT (MS) (MS) (MB/S) CONNER 210 19/5 8.33 1.5 CP-3200 MICROPOLIS 200 15/4 8.33 1-1.75 1770 (Zone Bit Recording) **** LIGHTING 320 12.5/1.5 6.89 2SUMMARY: Most SCSI drives announced in the exhibit have a readahead buffer in their embedded controller. Some venders have less SCSI overheads than current Maxtor's drive (827 us), but probably still in the range of 600 us. Maxtor will use 16-bit National HPC controller in their next generation controller. We can expect that the total SCSI overheads will drop to the range of 200 us by then. The disk industry will follow this direction to fix current SCSI number 1 problem, i.e. excessive bus overheads and command decode overhead **.
CAPACITY AVG SEEK/TRK-TRK LATENCY RAW DATA RATE (MB) FMT (MS) (MS) (MB/S) Century Data: 600 15/3.5 8.33 12.3 (5 modified SMD channels, 8in) Fujitsu Eagle 1000 16/4 8.25 18 (6 modified ESMD channels, 10.5in) IBIS 2800 16/2.5 8.33 12 (2 proprietary channels, 14in, VME bus connected) 4. Optical disks3M/Sony agreed to have a ISO standard (130 MM) read/write medium. Most other R/W optical manufactures probably will join this standard. This definitely has very positive impact to r/w optical disk industry. The WORM (write once) disk industry does not have a standard now. The r/w optical disk form factor probably will change to 86 mm in 1991.
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: Fear the Internet, was Cool Things You Can Do in z/OS Newsgroups: bit.listserv.ibm-main Date: 13 Apr 2011 13:45:38 -0700mike.a.schwab@GMAIL.COM (Mike Schwab) writes:
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: Fear the Internet, was Cool Things You Can Do in z/OS Newsgroups: bit.listserv.ibm-main Date: 13 Apr 2011 15:10:29 -0700scott_j_ford@YAHOO.COM (Scott Ford) writes:
mentions:
there was xmas exec on bitnet in nov87 ... vmshare archive
http://vm.marist.edu/~vmshare/browse.cgi?fn=CHRISTMA&ft=PROB
and was almost exactly a year before (internet) morris worm (nov88)
https://en.wikipedia.org/wiki/Morris_worm
xmas exec was social engineering ... similar to some current exploits which advertise something that victim has to download and then (manually) explicitly execute (requires victim's cooperation).
some additional
https://www.garlic.com/~lynn/2011b.html#10
misc. past posts mentioning bitnet (&/or earn)
https://www.garlic.com/~lynn/subnetwork.html#bitnet
which used technology similar to the corporate internal network (larger
than arpanet/internet from just about the beginning until late '85 or
early '86):
https://www.garlic.com/~lynn/subnetwork.html#internalnet
... and
https://www.garlic.com/~lynn/2011f.html#23
"fix" previous reference (missing trailing "l"):
https://web.archive.org/web/20090117083033/http://www.nsa.gov/research/selinux/list-archive/0409/8362.shtml
in '89 there was several messages that were sent out that major internal corporate (MVS-based) administrative systems had fallen victim to a virus ... however after several iterations it was eventually announced that the systems were suffering from some bug.
this selection of some internet related items/posts starts out with
reference to corporate installed email gateway in fall of '82
https://www.garlic.com/~lynn/internet.htm
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: Fear the Internet, was Cool Things You Can Do in z/OS Newsgroups: bit.listserv.ibm-main Date: 13 Apr 2011 16:08:01 -0700scott_j_ford@YAHOO.COM (Scott Ford) writes:
big difference between internal network and the internet in the 80s ... was that all internal network links (that left corporate premise) had to be encrypted. could be a big pain ... especially when links crossed certain national boundaries. in the mid-80s, it was claimed that the internal network had over half of all link encryptors in the world.
company also did custem encrypting PC (2400 baud) modems for corporate
home terminal program. there is folklore that one high-ranking (EE
graduate) executive was setting up his own installation at
home. supposedly at one point he stuck his tongue in rj11 jack (to see
if their was any juice ... old EE trick) ... just as the phone
rang. After that there was a corporate edict that all modems made by the
company had to have the jack contacts recessed sufficiently so babies
(and executives) couldn't touch them with their tongue.
https://en.wikipedia.org/wiki/RJ11
i had an HSDT (high-speed data transport) project and was dealing with
T1 links & higher speed. T1 link encryptors were really expensive
... but you could get them ... but I to start work on my own to go
significantly faster. misc. past posts mentioning HSDT
https://www.garlic.com/~lynn/subnetwork.html#hsdt
big difference in worms/viruses (and various other exploits) and social engineering ... is that social engineering requires active participation by the victim (current flavors frequently advertise download & execute things, frequently of very dubious nature; games, videos, etc). allowing users to execute arbitrary (unvetted) programs was identified as vulnerability at least back in the 70s (if not the 60s).
somewhat more recent thread (with some of my comments copied from
another venue)
https://www.garlic.com/~lynn/2011f.html#14 How is SSL hopelessly broken? Let us count the ways
How is SSL hopelessly broken? Let us count the ways; Blunders expose
huge cracks in net's trust foundation
http://www.theregister.co.uk/2011/04/11/state_of_ssl_analysis/
with regard to above ....
we had been called in to consult with small client/server startup that wanted to do payment transactions on their server, they had also invented this technology called SSL they wanted to us (the result is now frequently called "electronic commerce"). By the time we were finished with the deployments ... most of the (current) issues were very evident.
very early in the process I had coined the term "comfort certificate"
... since the digital certificate actually created more problems than it
solved ... in fact, in many cases, it was totally redundant and
superfluous ... and existed somewhat as magic "pixie dust" ... lots of
old posts mentioning SSL digital certificates:
https://www.garlic.com/~lynn/subpubkey.html#sslcert
There was two parts of SSL deployment for "electronic commerce" ... between the browser and webserver and between the webserver and the "payment gateway" ... I had absolute authority over interface deployment involving "payment gateway" ... but had only advisery over browser/webserver. There were a number of fundamental assumptions related to SSL for browser/webserver secure deployment ... that were almost immediately violated by merchant webservers (in large part because of the high overhead of SSL cut their throughput by 85-90%). I had mandated mutual authentication for webserver/gateway (implementation didn't exist originally) and by the time deployment was done the use of SSL digital certificates was purely a side-effect of the crypto library being used.
the primary use of SSL in the world today is electronic commerce for hiding payment transaction information. The underlying problem is the transaction information is dual-use both authentication and needed by dozens of business processes in millions of of places around the world. In the X9A10 financial working group we (later) directly addressed the dual-use problem with the x9.59 financial standard (directly addressing the problem, x9.59 is significantly lighter weight than SSL as well as KISS). This eliminated the need to hide the transaction details ... also eliminates the threat from majority of data breaches (doesn't eliminate data breaches, just eliminates crooks being able to use the information for fraudulent purposes). Problem (as always) is there is significant vested interests in the current status quo.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: First 5.25in 1GB drive? Newsgroups: alt.folklore.computers Date: Wed, 13 Apr 2011 19:32:51 -0400re:
found in old email log ... also just did search and found thread in
google usenet archive
Date: 16 Nov 89 07:46:10 GMT
Newsgroups: comp.periphs
Subject: Re: Need info about CDC Wren hard disks
Summary: a summary of interesting imprimis offerings
this is a brief summary of the latest offerings from imprimis.
as always, i take care but no responsibility for details; call
your imprimis rep. IN PARTICULAR, the prices are for me as an at&t
person; we get huge volume discounts, you should probably add 50-75%
to get a more available price.
the data comes from data sheets and salespeople. caveat empor.
wren vii:
the latest in a long line of disks. 5.25in 1.2GB SCSI disk.
average seek 16.5ms, 40KH MTBF. sustained 1.7MB/s.
available now as evaluation units at $3294, probable eventual cost
of ~$2500 ($2/MB).
elite:
new range of 5.25in disks (eventually replacing the wren's).
comes in SMD, IPI-2 and SCSI-2 interfaces. capacity is
1.2GB (1.5GB scsi). latency is <6ms, average seek 12ms.
sustained transfer of 3MB/s. 100KH MTBF.
smd evaluation units in jan, scsi production in apr/may
($4.5-5K).
sabre 2hp:
new version of the regular 8in sabre; 1.2GB and 50KH MTBF.
IPI-2 interface, sustained 6MB/s (twice regular sabres).
latency 8.3ms, ave seek 15ms.
these are shipping now, $7.4K.
sabre 2500:
2.5GB, evaluation jan/feb, seek 13ms, MTBF 100KH,
3MB/s, $8K.
arraymaster 9058:
(base for imprimis's raid).
this controller ($15K, at beta sites now) connects to drives
(any kind, speed) via IPI-2 and connects to a host via IPI-3.
assuming fast drives like sabre 2hp, host data rates are
25MB/s peak, 22MB/s sustained. imprimis will be selling a couple
of packages based on this controller; a small pseduo-disk of
5GB, 20MB/s sustained transfer, and a larger disk 16GB, with
two 18MB/s sustained i/o ports. both these packages have a lot
of internal error correction, a mean time to data loss of ~114yrs.
P.S. i note in passing that the WREN V and WREN VI were plagued with early firmw
problems regarding bus timeouts on long I/O transfers. these have been fixed
(my drives were fixed under warranty) and new drives should be okay.
but be wary of older drives.
... snip ...
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: First 5.25in 1GB drive? Newsgroups: alt.folklore.computers Date: Wed, 13 Apr 2011 20:08:20 -0400re:
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 14 Apr, 2011 Subject: US military spending has increased 81% since 2001 Blog: FacebookUS military spending has increased 81% since 2001; adjusted for inflation
some more detail:
http://www.cdi.org/program/document.cfm?documentid=4623
note cdi.org has moved to
http://www.pogo.org/straus/
... missing $1T is now
http://www.pogo.org/straus/issues/defense-budget/2010/what-did-the-rumsfeld-gates-pentagon-do-with-1-trillion.html
note cdi.org has moved to
http://www.pogo.org/straus/
... missing $1T is now
http://www.pogo.org/straus/issues/defense-budget/2010/what-did-the-rumsfeld-gates-pentagon-do-with-1-trillion.html
Winslow (no relation) may be just saying that Gates may be portrayed
that way if nothing substantive happens (which may be out of Gates
control). more detailed analysis (winslow one of the authors)
http://chuckspinney.blogspot.com/p/pentagon-labyrinth.html
... written by many of Boyd's "acolytes" (in the past, I had sponsored Boyd's briefings at IBM)
Gates got some points for referencing Boyd (a number of times) in this
speech:
http://www.defense.gov/speeches/speech.aspx?speechid=1443
I have way too many posts and references to Boyd here:
https://www.garlic.com/~lynn/subboyd.html
thread about military leaking into commercial
http://lnkd.in/6Kefvg
... by one of the labyrinth authors:
http://chuckspinney.blogspot.com/2011/02/why-boeing-is-imploding.html
... CSPAN had broadcast interview with several of the authors ... where the 787 scenario was highlighted
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: TCP/IP Available on MVS When? Newsgroups: bit.listserv.ibm-main Date: 14 Apr 2011 08:45:07 -0700Steve_Conway@AO.USCOURTS.GOV (Steve Conway) writes:
vmshare reference to 5798-fal:
http://vm.marist.edu/~vmshare/browse.cgi?fn=TCPIP&ft=PROB
the base implementation was later made available on MVS by moving over the VM code and writing a simulation for some of the VM functions.
later there was a contract to implement tcp/ip support in VTAM. the folklore is that when the implementation was first demo'ed ... the company said that it was only paying for a "correct" implementation ... and everybody knows that a "correct" tcp/ip implementation is significantly slower than LU6.2 (not significantly faster). The contract was handled by "local" ibm office in Palo Alto Sq office bldg.
now predating products ... there was various univ. implementations
... reference to tcp/ip at UCLA MVS in late 70s:
https://en.wikipedia.org/wiki/Bob_Braden
predates the great switchover from host/imp protocol to tcp/ip protocol on 1jan83.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: TCP/IP Available on MVS When? Newsgroups: bit.listserv.ibm-main Date: 14 Apr 2011 12:24:21 -0700lynn@GARLIC.COM (Anne & Lynn Wheeler) writes:
part of the issue was that the base support shipped with a box that was basically a channel-attached bridge (similar but different to 3174 boxes that supported LANs) ... so the host stuff had to do all the ARP/MAC/LAN layer gorp ... rfc1044 support was channel attached tcpip-router ... so a whole protocol layer was eliminated from host processing.
well before the tcp/ip support in vtam ... there was considerable misinformation regarding sna/vtam flying about ... including being useable for NSFNET backbone (operational precursor to the modern internet).
recent references to SNA/VTAM misinformation from the period:
https://www.garlic.com/~lynn/2011.html#4 Is email dead? What do you think?
https://www.garlic.com/~lynn/2011b.html#65 If IBM Hadn't Bet the Company
https://www.garlic.com/~lynn/2011c.html#12 If IBM Hadn't Bet the Company
https://www.garlic.com/~lynn/2011d.html#4 Is email dead? What do you think?
https://www.garlic.com/~lynn/2011e.html#25 Multiple Virtual Memory
https://www.garlic.com/~lynn/2011e.html#32 SNA/VTAM Misinformation
https://www.garlic.com/~lynn/2011e.html#34 SNA/VTAM Misinformation
https://www.garlic.com/~lynn/2011e.html#43 junking CKD; was "Social Security Confronts IT Obsolescence"
https://www.garlic.com/~lynn/2011e.html#57 SNA/VTAM Misinformation
https://www.garlic.com/~lynn/2011e.html#77 Internet pioneer Paul Baran
https://www.garlic.com/~lynn/2011e.html#83 History of APL -- Software Preservation Group
misc. old email regarding working with NSF on various activities
leading up to NSFNET backbone:
https://www.garlic.com/~lynn/lhwemail.html#nsfnet
This has postings regarding various announcements
http://vm.marist.edu/~vmshare/browse.cgi?fn=IBMNEW89&ft=MEMO
from above posted 9/21/84:
VM Interface Program for TCP/IP (5798-DRG): Provides VM user the
capability of participating in a network with TCP/IP transmission
protocol. Includes ability to do file transfers, send mail, and log on
remotely to VM hosts. (Comment: It's not clear whether this equals VM
access to non-VM hosts such as are found on ARPANET. I believe this is
the same product as WISCNET, already available to academic shops.)
... snip ...
and from above posted 4/22/87:
IBM also announced the new TCP/IP facility (5798-FAL) on 4/21/87.
This package replaces the old program (5798-DRG) and includes some
programs for PCs. The announcement is 287-165. To quote: "... IBM
TCP/IP for VM provides the VM/SP, VM/SP HPO, or VM/XA SF user with the
capability of participating in a multi-vendor Internet network using
the TCP/IP protocol set. This protocol set is an implementation of
several of the standard protocols defined for the Defense Advanced
Research Projects Agency. The use of these protocols allows a VM user
to interface with other systems that have implemented the TCP/IP
protocols. This connectivity includes the ability to transfer files,
send mail, and log on to a remote host in a network of different
systems. The IBM TCP/IP for VM program uses a System/370 channel
attached to a variety of controllers or devices for connection to the
selected network. The network protocols supported are IBM Token-Ring,
Ethernet(1) LAN, ProNET(2) and DDN X.25. IBM TCP/IP for VM offers IBM
TCP/IP for the PC as an optional feature, allowing the user of an IBM
personal computer on an IBM Token-Ring or Ethernet LAN to communicate
with the VM system using the TCP/IP protocols." Announced devices
supported are the IBM Series/1, 7170 DACU, and 9370 LAN adapters
(Token Ring or Lan)
... snip ...
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: TCP/IP Available on MVS When? Newsgroups: bit.listserv.ibm-main Date: 15 Apr 2011 07:48:43 -0700lynn@GARLIC.COM (Anne & Lynn Wheeler) writes:
additional trivia drift ... this is html version of internal IOS3270
"green card" (material from GX20-1850-3 and other sources)
https://www.garlic.com/~lynn/gcard.html
IOS3270 was cms application used for many applications ... including all
the service panels in the (4361/cms) 3090 "service processor" ... old
email reference
https://www.garlic.com/~lynn/2010e.html#email861031
above is with respect to including a failure analysis/debug tool (I had written) as part of 3092.
there had been a "blue card" for the 360/67 ... that included details of 67 features (like virtual memory support, multiprocessor control register detail) ... it also included sense data for several devices. I had provided the GCARD author with the sense data information. I had also included the sense data for the channel-attached tcp/ip router.
I still have a "blue card" in a box someplace ... obtained at the
science center from a fellow member (his name is "stamped" on the card).
GML had been invented at the science center in 1969 ... GML actually is
the first letter of the last names of the inventors (one of which is
"stamped" on my blue gard). A decade later, in the late 70s, GML becomes
an ISO standard as SGML.
https://www.garlic.com/~lynn/submain.html#sgml
Another decade & SGML morphs into HTML:
http://infomesh.net/html/history/early/
and the first webserver outside CERN, is on the SLAC VM/CMS system:
https://ahro.slac.stanford.edu/wwwslac-exhibit
part of the significant sna/vtam misinformation campaign in the late 80s
was to get the internal network converted over to sna/vtam. The campaign
to convert the internal network backbone to sna/vtam was so thick that
the backbone meetings were being restricted to management only (and
technical people were being excluded) ... recently posted old email:
https://www.garlic.com/~lynn/2011.html#email870306
in this post (linkedin Greater IBM discussion)
https://www.garlic.com/~lynn/2011.html#4 Is Email dead? What do you think?
In my HSDT effort I was doing multiple high-speed links and needed
sustained aggregate thruput in excess of channel speeds (load spread
across multiple mainframe channels).
https://www.garlic.com/~lynn/subnetwork.html#hsdt
Internal network VNET/RSCS used vm spool file system for intermediate storage. The API was synchronous and on heavily loaded system could limit thruput to 4-6 4kbyte blocks per second. I needed at least 100 times that thruput and had done a new API and (vm spool file) infrastructure to support sustained aggregate link thruput equivalent to multiple channels.
The internal network had been larger than arpanet/internet from just
about the beginning until late '85 or early '86.
https://www.garlic.com/~lynn/subnetwork.html#internalnet
and similar technology was used for BITNET/EARN (where this mailing list
originated).
https://www.garlic.com/~lynn/subnetwork.html#bitnet
Rather than converting the internal network links to SNA/VTAM, it would have been sigificantly better, cost/effective, and efficient to have converted the internal network links to tcp/ip.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 15 Apr, 2011 Subject: At least two decades back, some gurus predicted that mainframes would disappear Blog: Greater IBMre:
re: 747/cessnas ... the mega-datacenters ... is possible analogous to operating millions of bullet trains with a couple dozen operators.
a little more ("vax-killer") 9370 trivia from old email in 87:
https://www.garlic.com/~lynn/2011e.html#30
Recent reference to having worked on doing 14% thruput improvement on
450K+ statement cobol application that ran overnight batch. There were
40-some bloated CECs sized for the overnight batch window (@$30+m
per). They were doing frequent hardware upgrades with nothing older
than 18months.
https://www.garlic.com/~lynn/2011c.html#35
In 1980, (internal) STL (development lab) was bursting at the seams and they were in the process of moving 300 people from the IMS group to remote bldg ... with 3270 terminals connected back to STL datacenter. They had tried "remote" 3270 and found the human factors absolutely deplorable. I volunteered to write support for non-IBM channel-extender that allowed "local" 3270s to operate (at the remote bldg) connected back to the STL mainframes.
One of the things I did was simulate channel-check for certain kinds of unrecoverable transmission errors (channel extender operation included microwave telco transmission segment) ... which were fall through standard operating system retry operations. The vendor then tried to (unsuccessfully) talk IBM into allowing my support be shipped to customers. Finally they re-implemented the same design from scratch.
Role forward to the first year of 3090 customer installs and I'm contacted by 3090 product manager. There is a service that collects customer mainframe EREP (error reporting) data and generates regular summary reports of the information. The 3090 was designed to have an aggregate total of 5-6 channel check errors (in aggregate across all customer 3090s for period of year). It turns out that an aggregate of 20 channel check errors had been reported across all customers during the first year of operation ... the additional errors coming from the channel-extender software (simulating channel check error for unrecoverable transmission errors). I was asked if I could get the vendor to change their software emulation so it wouldn't simulate channel checks. After some amount of research, I determined that simulating IFCC (interface control check) would effectively result in the same EREP processing (w/o messing up the 3090 channel check reports).
at this NASA dependable computing meeting
https://web.archive.org/web/20011004023230/http://www.hdcc.cs.cmu.edu/may01/index.html
I ask the other vendors if any of them have reports of every error on every installed machine (and/or are they designed to have less than aggregate of dozen specific kinds of errors across all installed machines for period of year). Part of high integrity infrastructure starts with detailed instrumentation and corresponding reports ... in order to actually know what is going on.
Note that field engineering has had bootstrap, incremental diagnostic process that starts with scoping part in field. Starting with 3081 and TCMs, it was no longer possible to probe much of the hardware (inside TCM). As a result, "service" processor was introduced ... that the field engineer could "scope" the service processor ... and then use the "service processor" to diagnose the rest of the machine. For 3090, the "service processor" was initially "upgraded" to being a 4331 running a highly customized version of vm370/cms. Later the 3090 "service processor" (aka 3092) was upgraded to a pair of (redundant) 4361s (both running highly customized version of vm370/cms).
Lots of today's infrastructure involve network operation with large number of components using DBMS transaction semantics ... and network environment includes transmission retry. In such an environment ... it is possible to have five-nines high availability with redundant/fall-over systems ... where individual components just need to be designed to fail-safely.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 16 Apr, 2011 Subject: At least two decades back, some gurus predicted that mainframes would disappear Blog: Greater IBMre:
We complained long & loud that the 3274/3278 was much worse for interactive computing than the 3272/3277 it was replacing. Eventually we got an answer back that 3274/3278 wasn't targeted for interactive computing, it was targeted for data entry.
Late 80s, a senior disk engineer got a talk scheduled at the annual,
world-wide internal communication group conference ... and opened the
talk with statement that the communication group was going to be
responsible for the demise of the disk division. The issue was that
the communication group had a stranglehold on the datacenter and also
had strategic ownership of products that "crossed" the datacenter
walls. In disk sales numbers, the disk division saw leading edge of
tsunami data fleeing the datacenter for more distributed computing
friendly platforms. The disk division had come up with several
products to address the problem, but were constantly being vetoed by
the communication group (since they had strategic ownership for
products that involved crossing datacenter walls). misc. past posts
mentioning stranglehold the communication group had on the datacenter
https://www.garlic.com/~lynn/subnetwork.html#emulation
In the late 80s, while the communication group was blocking datacenter
involvement in client/server, we had come up with 3-tier network
architecture (middle layer), was including it in gov. request
responses and out pitching it to customer executives (and were taking
constant barbs in the back from the communication group)
... misc. past posts
https://www.garlic.com/~lynn/subnetwork.html#3tier
The original SQL/relational implementation was System/R in bldg. 28. There use to be "friendly" arguments between IMS down in STL and System/R. The ("physical") IMS group argued that relational doubled the disk space requirements (for the "implicit" indexes) ... and also greatly increased the disk i/os (also for reading the "implicit" indexes). The System/R group countered that exposing the explicit record pointers (as part of the data) greatly increased manual maintenance.
In the 80s, there was significant increase in system real storage
sizes (allowing caching of the relational indexes, significantly
reducing relational i/o overhead penalty) and significant decrease in
the disk cost per byte (significantly reducing index cost penalty). At
the same time people expense was going up and number of skilled people
failing to keep up with the customer appetite for DBMS. SQL/Relational
reduced the people time&skill for DBMS support (compared to 60s
"physical" DBMS). misc past posts mentioning original relational/SQL
https://www.garlic.com/~lynn/submain.html#systemr
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 16 Apr, 2011 Subject: Early mainframe tcp/ip support (from ibm-main mailing list) Blog: Old Geek Registryre:
We had been working with director of NSF for some time ... loosely
around a project I called HSDT ... with many of the sites that were to
become part of the NSFNET backbone ... and were meeting lots of
internal opposition. At one point, NSF had budget proposal for HSDT to
get $20M ... but congress cut back on budget for such activity. Then
the internal opposition really kicked in (along with lots of SNA/VTAM
misinformation in many areas) ... and we weren't allowed to
participate in the actual NSFNET backbone at all. Misc. past email
https://www.garlic.com/~lynn/lhwemail.html#nsfnet
The NSFNET T1 backbone RFP specified "T1" (in large part because I already had T1 up and running production). The winning bid actually put in 440kbit links and then possibly to try and meet the letter of the RFP ... but in T1 trunks with telco multiplexors managing multiple 440kbit links over the T1 trunks (some sniping about why they couldn't call it a T5 network ... since some of the T1 trunks may have been in turn, multiplexed over T5 trunks).
Then when there was the NSFNET-II backbone (upgrade to T3), I was
asked to be the red team (possibly anticipating that they would be
able to shutdown my sniping) and there were a couple dozen people from
a half-dozen labs around the world for the blue team. At the final
review, I presented first. Then 5mins into the blue team presentation,
the person running the review ... pounded on the table and said that
he would lay down in front of a garbage truck before he allowed any
but the blue team proposal to go forth. misc. past posts mentioning
NSFNET activity
https://www.garlic.com/~lynn/subnetwork.html#nsfnet
semi-related recent post (mentioning talk in the late 80s about the
communication group was going to be responsible for the demise of the
disk division):
https://www.garlic.com/~lynn/2011f.html#33
in this Mainframe Experts discussion
http://lnkd.in/mk79ZS
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 16 Apr, 2011 Subject: At least two decades back, some gurus predicted that mainframes would disappear Blog: Greater IBMre:
Possibly because GAO didn't believe SEC was doing anything ... they started doing reports of public company financial filings that they felt were fraudulent and/or significant accounting errors ... there was even uptick after SOX. Apparently the motivation was boosting executive bonuses ... and even if the filings were later corrected, the bonuses weren't adjusted.
http://www.gao.gov/new.items/d061053r.pdf ,
https://www.gao.gov/products/gao-06-1079sp
recently from somewhere on the web: "Enron was a dry run and it worked so well it has become institutionalized"
a somewhat similar picture was painted in congressional hearings into Madoff ponzi scheme by the person that had tried for a decade to get SEC to do something about Madoff.
and then Cramer's comments whether SEC would ever do anything about
illegal naked short selling
http://nypost.com/2007/03/20/cramer-reveals-a-bit-too-much/
SOX also required SEC to do something about the rating agencies
... that played pivotal role in the recent financial mess ... but
nothing appears to have been done except report:
http://www.sec.gov/news/studies/credratingreport0103.pdf
some of what actually went on came out in fall2008 congressional
testimony ... but there appeared to be little followup ... although
just recently there is this:
http://www.computerworlduk.com/news/it-business/3274277/bank-email-archives-thrown-open-in-devastating-financial-crash-report/1
possibly attempting to explain little or nothing being done:
http://blogs.forbes.com/neilweinberg/2011/04/14/corrupt-bank-oversight-is-creating-new-immoral-hazard/
http://www.nytimes.com/2011/04/14/business/14prosecute.html?_r=2&hp
minor nit ... austin romp (801) was originally targeted as displaywriter follow-on ... when that was killed, the box was retargeted for the unix workstation market ... and the company that had done the AT&T unix port for ibm/pc pc/ix was hired to do "AIX" port (release 2) along with PC/RT. That morphed into AIX (release 3) for power/rios.
Somewhat in parallel, the palo alto group had been working with UCLA Locus system (unix work-alike that supported both distributed file system and distributed process execution). The palo alto group ported UCLA Locus to both mainframe and 386 ... and was released as product called aix/370 and aix/386.
People in Palo Alto had also been working with Berkeley with their UNIX work-alike (BSD) ... which was originally going to be ported to mainframe and released. However, they got redirected to do the port to the PC/RT instead and it was released as "AOS" (as a PC/RT alternative to AIX).
There was also a number of groups working with CMU on their AFS (distributed filesystem) and MACH (CMU's unix work-alike). A number of companies picked up MACH for their "unix" system ... including Jobs at NeXT. When Jobs returned to Apple, be brought MACH along and it morphed into the (new) Apple (MAC) operating system.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 16 Apr, 2011 Subject: Early mainframe tcp/ip support (from ibm-main mailing list) Blog: Old Geek Registryre:
recent post I did on facebook:
FTP is 40 years old
http://www.bit-tech.net/news/hardware/2011/04/15/ftp-is-40-years-old/1
FTP still forms the backbone of many parts of the Internet, despite
its age. The backbone of the Internet, FTP (file transfer protocol),
celebrates its 40th birthday tomorrow. Originally launched as the RFC
114 specification, which was published on 16 April 1971, FTP is
arguably even more important today than when it was born.
... snip ...
one of my responses to some of the comments:
File Transfer Protocol (rfc959) talks about the history (sec2.1)
... as well as some issues/problems caused by remapping to TCP
(including needing separate control & data streams, rfc765). who does
separate control&data streams on tcp these days?
... snip ...
aka arpanet host/imp protocol provided for an "out-of-band" control mechanism ... which wasn't carried over into tcp/ip ... so remapping FTP to TCP/IP they added a separate session for the control traffic.
for other references ... my IETF RFC index
https://www.garlic.com/~lynn/rfcietff.htm
at interop '88 (tcp/ip interoperability in 1988), there was
significant number of OSI (various x.nnn stuff) in various booths
... probably because the gov. had mandated the elimination of tcp/ip
and change-over to gosip. I had a pc/rt in a booth ... but not IBM's
... it was at right angle adjacent to the SUN booth (in the central
isle) and Case/SNMP (still had not won the monitor wars) was in the
SUN booth. Case was talked into coming over and doing SNMP
build/install on the PC/RT. misc. past posts mentioning Interop '88
https://www.garlic.com/~lynn/subnetwork.html#interop88
as mentioned there was huge SNA/VTAM misinformation at the time
... part of it to convert the internal network to SNA/VTAM
(bitnet/earn was using similar technology to the internal network)
... including lots of stuff that SNA/VTAM would even be applicable to
NSFNET backbone. Reference to significant amount of email that had
been collected to SNA/VTAM misinformation:
https://www.garlic.com/~lynn/2006w.html#email870109
other references to SNA/VTAM misinformation ... for the internal
network ... they had change the internal backbone meetings to exclude
the technical people and have managers only
https://www.garlic.com/~lynn/2006x.html#email870302 ,
https://www.garlic.com/~lynn/2011.html#email870306
for other drift ... old email about getting EARN going:
https://www.garlic.com/~lynn/2001h.html#email840320
part of the above was because I had worked with the person on cp67 in
the early 70s. then I was blamed for online computer conferencing on
the internal network in the late 70s and early 80s (folklore was that
when the executive committee was told about online computer
conferencing and the internal network; 5of6 wanted to fire me
immediately). Part of the activity was referred to as Tandem Memos
... from IBM Jargon:
Tandem Memos - n. Something constructive but hard to control; a fresh
of breath air (sic). That's another Tandem Memos. A phrase to worry
middle management. It refers to the computer-based conference (widely
distributed in 1981) in which many technical personnel expressed
dissatisfaction with the tools available to them at that time, and
also constructively criticized the way products were [are]
developed. The memos are required reading for anyone with a serious
interest in quality products.
... snip ...
some rumors about Tandem Memos even leaked outside IBM and there was article in Nov81 Datamation.
Note one of the differences between IETF (tcp/ip) and ISO (OSI) was that IETF required different interoperable implementations before standards progression ... while ISO didn't require that a standard even be implementable before passing as standard.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 17 Apr, 2011 Subject: At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened Blog: Mainframe Expertsre:
also:
http://lnkd.in/mk79ZS
"VM IS DEAD" tome from 1989 (from employee involved in development)
containing some of the items that contributed to view that mainframe
was dying
https://www.garlic.com/~lynn/2011e.html#95
vm/4300 sold head-to-head against vax/vms in the mid-range market over
the whole period in about similar numbers ... in orders involved small
number of machines ... place where 43xx outdid vax/vms was in large
corporate orders involving multiple hundred of machines. The large
43xx orders were being placed out in every corporate nook&cranny
... internally it contributed to dept. conference rooms becoming
scarce commodity (so many being taken over for 43xx dept. machines).
It was the leading wave of distributed computing. Old 43xx email (back
to before 43xx announced & shipped)
https://www.garlic.com/~lynn/lhwemail.html#43xx
By the mid-80s, the mid-range market was starting to be taken over by
workstations and large PC. The 4331/4341 follow-ons (4361 & 4381) were
expected to see similar large sales ... which never materialized. This
is decade of vax numbers sliced & diced by year, model, us/world-wide
... showing the big plunge in the mid-80s (something similar was seen
by 43xx):
https://www.garlic.com/~lynn/2002f.html#0
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: VM IS DEAD tome from 1989 Newsgroups: alt.folklore.computers Date: Sun, 17 Apr 2011 17:22:22 -0400VM IS DEAD
and something similar in thread about predictions from early 90s
that the mainframe was dead
https://www.garlic.com/~lynn/2011f.html#37
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 17 Apr, 2011 Subject: At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened Blog: Mainframe Expertsre:
The "VM IS DEAD" tome (ala vm370) was written in 1989 when it appeared to be on its death bed. All sorts of data was fleeing the datacenter (given the stranglehold that the communication group had on the datacenter) ... behind the talk by senior disk engineer gave about the communication group was going to be responsible for the demise of the disk division ... also behind the rumors in the early 90s about the death of the mainframe.
In the 90s, a few operations found that they couldn't easily move some kinds of data off the mainframe which allowed it to linger on ... and around the start of the century, virtual linux breathed a whole new life into VM.
In the wake of the failed Future System project ... there was mad rush
to do a number of things ... XA-architecture (code name of "811" for
nov78 date on many of the documents), 303x (3031 was 158-3 with new
covers, 3032 was 168-3 with new covers and 3033 which started out as
168-3 logic/wiring remapped to faster chips), and 3081 (which would
implement some of "811"). Some details about FS, XA-architecture,
303x, 3081, and "811" here:
http://www.jfsowa.com/computer/memo125.htm
and
https://people.computing.clemson.edu/~mark/fs.html
Head of POK had managed to convince corporate to kill off VM in the mid-70s, shutdown the development group and move all the people to POK as part of MVS/XA development (or otherwise MVS/XA would miss its ship scheduled). There was joke about head of POK being major contributor to VAX/VMS since so many people from the development group left and went to work on VMS (rather than move to POK). Endicott managed to save the VM product mission, but had to reconstitute a development group from scratch.
The VM thing in 3081/XA was SIE ... but it was never intended for customer use ... purely for internal only tool supporting MVS/XA development. Because of limited microcode space in 3081, SIE instruction microcode was "paged" ... giving it very bad performance (but performance wasn't an issue because it was originally purely for internal MVS/XA development and test).
SIE was significantly enhanced in 3090 and extended with PR/SM (the microcode base for LPARs) in response to Amdahl's "hypervisor" (and finally given "real" life).
There was a small amount of "811" retrofitted to 3033 by the person credited with Itanium (was at IBM at the time and then left to work for HP) ... called dual-address space mode ... which was specifically to address an enormously growing problem in MVS with the Common Segment (CSA) ... aka had nothing to do with VM.
Virtual machines were originally done by the science center on a
specially modified 360/40 (with virtual memory hardware support)
... and called cp/40. Then when an "official" 360 with virtual memory
hardware started shipping, the science center replaced its 360/40 with
360/67 and cp/40 morphed into cp/67 .... and the science center
started making it available to customers. Misc. past posts mentioning
the science center (4th flr, 545 tech sq)
https://www.garlic.com/~lynn/subtopic.html#545tech
The cp/67 group split off from the science center and took over the Boston Programming Center on the 3rd flr ... and started the morph from cp67 to vm370. When the vm370 group outgrew the 3rd flr ... they moved out into the old SBC bldg in burlington mall (SBC having been given to CDC as part of legal settlement). It was the burlington mall group that was shutdown by the head of POK in the mid-70s (with many of the people deciding to leave and go to work for DEC).
misc. past posts mentioning FS
https://www.garlic.com/~lynn/submain.html#futuresys
other posts in this thread:
https://www.garlic.com/~lynn/2011e.html#15 At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened
https://www.garlic.com/~lynn/2011e.html#16 At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened
https://www.garlic.com/~lynn/2011e.html#17 At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened
https://www.garlic.com/~lynn/2011e.html#19 At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened
https://www.garlic.com/~lynn/2011f.html#32 At least two decades back, some gurus predicted that mainframes would disappear
https://www.garlic.com/~lynn/2011f.html#33 At least two decades back, some gurus predicted that mainframes would disappear
https://www.garlic.com/~lynn/2011f.html#35 At least two decades back, some gurus predicted that mainframes would disappear
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: The first personal computer (PC) Newsgroups: alt.folklore.computers Date: Mon, 18 Apr 2011 09:22:05 -0400jmfbahciv <See.above@aol.com> writes:
there are some number of recent reports that evolution has attempted to optimize brain energy use ... since the brain is one of the major energy using organs. all sorts of brain activity is being described as accomplished with the minimum use of energy (which is attributed to be a survival characteristic) ... aka only using as much of the brain as necessary.
misc. recent posts:
https://www.garlic.com/~lynn/2010q.html#14 Compressing the OODA-Loop - Removing the D (and mayby even an O)
https://www.garlic.com/~lynn/2010q.html#61 Compressing the OODA-Loop - Removing the D (and maybe even an O)
https://www.garlic.com/~lynn/2011.html#7 Looking for a real Fortran-66 compatible PC compiler (CP/M or DOSor Windows
https://www.garlic.com/~lynn/2011.html#39 The FreeWill instruction
https://www.garlic.com/~lynn/2011.html#78 subscripti ng
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: CPU utilization/forecasting Newsgroups: bit.listserv.ibm-main Date: 18 Apr 2011 06:36:19 -0700dba@LISTS.DUDA.COM (David Andrews) writes:
the science center had started accumulating all its (virtual machine) system activity from the 60s ... and established it as standard process for normal operation. By the mid-70s, the science center not only had several years of its own data ... but was also acquiring similar data from large number of internal datacenters. This was used for a lot of modeling and simulation work, along with workload & configuration profiling ... which eventually evolves into capacity planning.
one of the science center's models (implemented in APL) ... was made available (starting in the mid-70s) on the (internal online virtual machine) HONE systems (providing world-wide sales & marketing support) as the Performance Predictor. Sales people could collect customer workload & configuration profile and ask "what-if" questions (of the Performance Predictor) about changes to workloads and/or configuration.
misc. past posts mentioning science center
https://www.garlic.com/~lynn/subtopic.html#545tech
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 18 Apr, 2011 Subject: At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened Blog: Mainframe Expertsre:
The 303x was pre-XA ... was quick&dirty mad rush to get something out as fast as possible after the death of FS ... with FS having killed off nearly all 370 activity ... viewing it as competition. 303x went on in parallel with "811" and 3081 ... see Sowa and Clemson references, 3081 was going to take 7-8yrs from start, there was nothing in the pipeline because FS had killed off all competition ... so something had to be done as fast as possible. FS killing off all (internal) competition is also attributable for giving clone processors foothold in the market.
FE had a bootstrap service protocol that starts with being able to scope components. TCMs were introduced with 3081 ... with components inside the TCM and no longer "scopeable". Service processor was introduced to address the FE process. The (primitive UC-based) service processor could be scoped & diagnosed ... and then the 3081 was built with all sorts of probes into the TCMs that were accessible by the service processor (FEs being able to use the service processor to diagnose the rest of the machine). However, UC-based service processor was really too primitive to do a whole lot.
The 3090 service processor (3092) started out as 4331 with FBA disk
running a highly modified version of VM370 release 6 (with FEs being
able to scope/diagnose the 4331 and then use the 3092 to diagnose the
3090 TCMs. By time 3090s started shipping, the "3092" had been
upgraded to a pair of redundant 4361s (running highly modified version
of VM370 release 6). It was this facility that was sophisticated
enough to handle PR/SM and emerging LPAR (to compete against Amdahl's
hypervisor). As an aside ... all the 3090 service processor screens
were done in CMS IOS3270 ... mentioned in this "Old Geek" discussion:
http://lnkd.in/UkVnRg
about early mainframe tcp/ip support (also ran in ibm-main mainframe
mailing list).
I was heavily involved with the 138/148 ... precursor to 4331/4341
... and then with 4331/4341 ... as previously mentioned misc. old
email about 43xx
https://www.garlic.com/~lynn/lhwemail.html#43xx
I had been con'ed into helping with parts of the 138/148 ... and then the endicott product manager (code name was virgil/tully) con'ed me into running around the world with him making presentations to the various country product forecasters. World Trade (non-US) forecasts tended to be taken at face value ... since the forecast went from the plant books and showed up on the country books for sale to customers. US forecasts had turned into whatever hdqtrs said was "strategic" (since machines went directly from the plant's books to the customer ... w/o ever showing on the books of the sales organization ... and therefor there was no downside for US sales to forecast whatever hdqtrs told them was strategic). There was eventually joke in the US that labeling something as "strategic" was method of pushing a product that customers weren't buying (since US "strategic" carried with it various kinds of sales incentives).
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 18 Apr, 2011 Subject: Massive Fraud, Common Crime, No Prosecutions Blog: FacebookMassive Fraud, Common Crime, No Prosecutions
There had been some securitized mortgages (CDOs) in the S&L crisis with doctored supporting documents for fraud. In the late 90s, we were asked to look at what could be done for integrity/assurance of securitized mortgages (CDOs) supporting documents.
In the first of this century, loan originators found they could pay rating agencies for triple-A ratings and immediately unload every loan (without regard to quality or borrower's qualifications). Speculators found that no-down, no-documentation, 1% interest payment only ARMs could make 2000% ROI buying&flipping properties in parts of the country with 20%-30% inflation.
Buyers of triple-A rated toxic CDOs didn't care about supporting documents, they were buying purely based on the triple-A rating. Rating agencies no longer cared about supporting documentation because they were being payed to give triple-A rating. Supporting documentation just slowed down loan originator's process of issuing loans. Since nobody cared about supporting documentations, they became superfluous ... which also resulted in there no longer being issue about supporting documentation integrity/assurance.
Lending money used to be about making profit on the loan payments over the life of the loan. With triple-A rated toxic CDOs, for many, it became purely a matter of the fees & commissions on the transactions and doing as many as possible. There were reportedly $27T in triple-A rated toxic CDO transactions done during the bubble ... with trillions in fees & commissions disappearing into various pockets.
Possibly aggregate 15%-20% take on the $27T ($5.4T) as the various transactions wander through the infrastructure (starting with original real estate sale).
In the fall2008 congressional hearings into the rating agencies ... the issue was raised that the rating agencies might blackmail the gov. into not prosecuting with the threat of downgrading the govs' credit rating (an issue that is in the news today).
there was report that wall street tripled in size (as percent of GDP) during the bubble and NY state comptroller reported that wall street bonuses spiked over 400% during the bubble ... all being fed by the $27T in triple-A rated toxic CDO transactions.
recent posts mentioning the $27T:
https://www.garlic.com/~lynn/2011.html#50 What do you think about fraud prevention in the governments?
https://www.garlic.com/~lynn/2011.html#80 Chinese and Indian Entrepreneurs Are Eating America's Lunch
https://www.garlic.com/~lynn/2011b.html#27 The Zippo Lighter theory of the financial crisis (or, who do we want to blame?)
https://www.garlic.com/~lynn/2011b.html#42 Productivity And Bubbles
https://www.garlic.com/~lynn/2011b.html#43 Productivity And Bubbles
https://www.garlic.com/~lynn/2011b.html#45 Productivity And Bubbles
https://www.garlic.com/~lynn/2011c.html#46 If IBM Hadn't Bet the Company
https://www.garlic.com/~lynn/2011d.html#23 The first personal computer (PC)
https://www.garlic.com/~lynn/2011e.html#7 I actually miss working at IBM
https://www.garlic.com/~lynn/2011e.html#36 On Protectionism
https://www.garlic.com/~lynn/2011e.html#48 On Protectionism
https://www.garlic.com/~lynn/2011e.html#60 In your opinon, what is the highest risk of financial fraud for a corporation ?
https://www.garlic.com/~lynn/2011e.html#74 The first personal computer (PC)
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: The first personal computer (PC) Newsgroups: alt.folklore.computers Date: Mon, 18 Apr 2011 14:21:54 -0400Peter Flass <Peter_Flass@Yahoo.com> writes:
increase by factor of ten could require ten times the blood flow and oxygen uptake ... as well as getting rid of the additional generated heat; likely would require some serious augmentation to pump in the extra blood/oxygen and remove the extra heat.
serious overclockers are using liquid nitrogen to go from 2.8 to 5.5
(about double):
http://www.liquidnitrogenoverclocking.com/index.shtml
http://www.tomshardware.com/reviews/5-ghz-project,731.html
would take some serious redoing how the brain was layed out and augmented.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 18 Apr, 2011 Subject: At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened Blog: Mainframe Expertsre:
IBM had parallel bus&tag that was 4.5mbytes. Cray channel was standardized as HIPPI parallel at 100mbyte/sec (big push for standard by LANL). IBM tried to support HIPPI on 3090 but 3090 couldn't handle the I/O rate. They did an ugly hack that cut into the side of the expanded storage bus and HIPPI transfers were done with peek/poke paradigm to special addresses.
In parallel with LANL & HIPPI, LLNL had serial interface that was being pushed as standard called FCS ... 1gbit/sec dual-simplex (i.e. concurrent 1gbit/sec in each direction).
In parallel with LANL & LLNL, somebody from SLAC was working on dual-simplex standard ... SCI ... with specification for both asynchronous memory bus operation and as I/O operation
There had been fiber technology knocking around POK for quite some time ... Austin (& Rochester) took the specification ... tweaked it to be 10% faster and use high reliable, commodity drivers and it shipped as SLA (serial link a adapter) dual-simplex ... simultaneous transfer in both directions. POK eventually ships the slower & much more expensive as ESCON ... but limited to transferring in only one direction at a time (simulating existing parallel IBM channels). In part because ESCON was still simulating half-duplex ... it still had thruput issues because of end-to-end turn-around latencies.
We had been working with LANL, LLNL and SLAC in all three standards bodies. The Austin engineer that did SLA ... started working on faster 800mbit version ... and we convince him to instead join the FCS standards work (where he takes over responsibility for editing the FCS standards document). This is crica 1990. Eventually some mainframe channel engineers start attending the FCS meetings attempting to do extremely unnatural things to the standard ... which eventually morphs into FICON (I still have much of the FCS standard mailing list distribution in archive some place).
Part of the hype for large number of half-duplex mainframe channels was problem with the 3880 controller and 3090s. Data streaming and 3mbyte transfers (for 3380 disk) was introduced with 3880 controller. The 3880 was much slower than 3830 ... so had special hardware path for data transfers ... leaving slow processor for handling commands & control path. At some point the 3090 engineers realize that the slow 3880 controller would have significantly higher channel busy and they needed to have a whole lot of additional channels (to offset the high 3880 channel busy overhead). The extra channels resulted in requiring extra manufacturing TCM. There was joke that 3090 product was going to bill the 3880 group for the increase in manufacturing costs for the extra channels & TCM.
Base SLA, FCS, and SCI operation was dual-simplex with asynchronous operation (not end-to-end serialization). Commands went down one path ... and that (channel) path was busy only for the duration of the transfer. The "channels" didn't have the half-duplex serialization busy overhead for control operations (with controllers & devices). As a result, asynchronous dual-simplex needed capacity for the raw transfers (but not the enormous amount of excess capacity to handle the half-duplex end-to-end latency for control operations).
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 19 Apr, 2011 Subject: At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened Blog: Mainframe Expertsre:
recent mainframe announcements includes support for out-of-order execution ... (announcement claims) contributing about 20% instruction throughput boost (this has been feature that has been seen in other processors for two decades or more).
the issue is that on a cache miss, the instruction stalls and waits for memory operation ... latency for the memory operation can be hundreds or even thousands of machine cycles. out-of-order operation allows following instructions to be executed asynchronously with the stalled instruction waiting for the cache miss memory operation (assuming high probability that they don't also have cache miss and/or dependent on the results from the stalled instruction).
the dual-simplex serial channels ... again from at least two decades ago ... did something similar ... but for I/O operations ... in high activity concurrent workload could have throughput improvements significantly higher (than the 20% for recent mainframe out-of-order execution ... as compared to the traditional mainframe half-duplex i/o paradigm)
note that a major reason for SSH in "811" architecture ("811" was mad rush to put together something after the failed Future System effort and supposedly comes from the Nov78 date on many of the documents) and introduced with 3081 ... was the significant I/O redrive serialization latency with operating system interrupt handlers. The issue was that there are multiple queued requests ... all managed by software ... the first queued operation was not started until after the interrupt handler finished processing the current active operation. This interrupt handling/redrive serialization latency was becoming an increasingly significant part of throughput.
SSH could have multiple queued requests (offloaded into hardware) ... and allow outboard processor to start the next request asynchronously with interrupt handling of the previous request (eliminating much of the I/O redrive latency). Dual-simplex serial channels tended to accomplish the same thing (in addition to enormously increasing utilization of the media bandwidth) ... but allowing command packets to be asynchronously transmitted down the outgoing path.
We had done something approx. similar to SSH for 2305 fixed-head disks with multiple exposures ... using the multiple exposures to queue up a lot of different requests ... which would execute asynchronously while processing interrupts from previous requests. This allowed achieving sustained throughput very close to theoretical maximum device transfer.
The san jose mainframe disk development (in bldg 14) had a variety of
mainframes for testing development disks. These mainframes were
running standalone with dedicated test time being scheduled 7x24. They
had once tried MVS to do concurrent testing in operating system
environment ... but found that with just a single "testcell", MVS had
15min. MTBF (requiring reboot). I offered to do an bulletproof IOS
rewrite that would never fail and support multiple, concurrent,
on-demand testing (significantly improving development
productivity). misc. past posts getting to play disk engineer in
bldgs. 14&15
https://www.garlic.com/~lynn/subtopic.html#disk
When "811" SSH was published ... I tried an absolutely optimized (production) interrupt-handler/IO-redrive pathlength (in software) to see how close I could come to SSH design (that used independent, outboard asynchronous processor). Turns out that SSH was largely justified based on extremely bloated operating system implementation with enormously long elapsed time between interrupt and I/O redrive.
One of the reasons that two decades ago, mainframes were being viewed as a plodding dinosaur ... already dead ... but the signal hadn't yet reached its brain ... was lots of vendors/products were doing asynchronous (dual-simplex) operations for latency masking/compensation. Base FCS standard two decades ago was asynchronous, dual-simplex 1gbit links ... being able to achieve nearly 2gbit aggregate thruput (concurrent operations in both directions). Something similar was seen in SCI standard ... with definitions for both asynchronous memory bus operation (between processor cache and memory) as well as I/O operations.
Even IBM had done something similar two decades ago in Hursley with Harrier/9333 ... which was basically SCSI disks ... with SCSI commands packetized and sent asynchronously down Harrier dual-simplex serial links. Harrier (for high concurrent workload environment) sustained significantly higher throughput than the same disks used in conventional half-duplex SCSI I/O operation.
I had tried to get Harrier to evolve into interoperability with FCS (being able to plug into FCS switches ... but operating at lower transfer rate) .... but instead it evolved into its own (incompatible) specification. Some of the mainframe channel engineers may have been involved ... since they were trying to co-op FCS and layer a bunch of unnatural things on top the base standard ... which evolves into FICON (interoperable harrier on same FCS switches with "FICON" may have made "FICON" look bad).
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: First 5.25in 1GB drive? Newsgroups: alt.folklore.computers Date: Tue, 19 Apr 2011 13:09:57 -0400re:
and some '91 history ... from long ago and far away
Date: 14 Feb 91 09:03:07 GMT
To: wheeler
Subject: Fujitsu Announces 2 GB, Ave. Positioning 11ms, 5.25in DASD
On February 14, 1991, Fujitsu announced and began OEM marketing four
models of M2652 series DASD, a new high-capacity 2GB, 5.25in DASD family.
This is the highest capacity available on a 5.25in devices from any
vendor.
The average access time is 11 msec. The average latency is 5.6 msec
(5,600 RPM). This was achieved by using thin film heads, thin film
disks and a light-weight actuator. The data transfer rates are 4.758
MB/sec (IPI-2), 10 MB/sec (SCSI Synchronous, W/256KB buffer) and 3
MB/sec (SCSI Asynchronous).
The M2652 will be shipped from June 1991. Samples will be priced at
1,000 KYen. Fujitsu plans to sell 500,000 units of these four models
(total) over three years.
Fujitsu 5.25-Inch HDD Specifications +-----------------------+-----------------------------+ | Model | M2652 Series | |-----------------------+-----------------------------| | Sample Price (Yen) | 1,000 KYen | | (dollar @ 130Y/$) | $7,692 | | Interface | IPI-2, SCSI | |-----------------------+-----------------------------| | Capacity | 2 GB | | Data Rate | 4.758 MB/sec | | - SCSI Synch(256KB BF)| 10 MB/sec | | - SCSI Asynch | 3 MB/sec | | Avg Positioning Time | 11 msec | | RPM | 5,400 | | Latency | 5.6 ms | |-----------------------+-----------------------------| | Bytes per Track | 52,864 | | Cylinders | 1,893 | | No of Disks | 12 | | No of Heads R/W + SR | 20 + 1 | | Dimension(IPI, SCSI-D)| 146mm x 220mm x 83mm | | - WxDxH (Others) | 146mm x 203mm x 83mm | | Weight | 4 Kg | | Power | +12V, +5V | +-----------------------+-----------------------------+
for random other piece of info ...
Date: 19 Mar 91 17:59:57 GMT
To: wheeler
Subject: PC OPERATING SYSTEMS, 1990
According to 3/15 NY Times "Business Day" (Source: Dataquest)
1990 unit shipments of personal computer operating systems in U.S.
were as follows:
SYSTEM QTY MS-DOS 14,021,000 (75%?) Windows 2,039,000 MacIntosh 1,946,000 UNIX 399,000 OS/2 300,000 (1.6%?)... snip ... top of post, old email index
a little more disk:
Date: 07 Aug 91 15:33:44 GMT
To: wheeler
Subject: Toshiba 3.5in 1-Gigabyte drive (Price=$2000)
"Toshiba begins US Output of one-Gigabyte Disk Drive: (Price = $2000)
"Toshiba Corp. said it has begun US Production of the first 3.5in
disk drives with memory capacity of one gigabyte, or roughly one
billion characters of information.
... snip ... top of post, old email index
and then there is
Date: 03 Sep 91 07:38:55 GMT
To: wheeler
Subject: Glass coated ceramic disks from KYOCERA, Japan
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com Date: 19 Apr, 2011 Subject: A brief history of CMS/XA, part 1 Blog: IBM Historic Computingalso posts to (linkedin) z/VM
aka "811" was name for architecture enhancements to 370
... started appearing with 3081:
Date: 07/16/80 19:42:07
From: wheeler
well, where do things stand ??? there has been no feedback on friday
meeting and/or current status. I did talk to XXXXX and he is putting
me on the distribution list for VM/811 specs. YYYYY (somebody in POK)
read all four or five papers I sent and wanted them classified "IBM
Confidential". I guess my management didn't understand initially about
such things and thot they had to go along. There is nothing in them
that warrants such a classification (& I know of several papers of the
same nature that were published w/o any classification). Hopefully
they will get approval in time for Atlanta Share meeting and can be
distributed at that time. CMS file system committee wants to see the
documentation on PAM. Baybunch had discussion and presentations about
VMCF/IUCV and service virtual machines. Longwait problem came up, fix
that ZZZZZ did to DMKDSP to check for outstanding SENDR requests
before setting long wait isn't going out. FE is closing APAR as things
currently work the way they are designed to. Tymshare also made
presentation on simple modification to CMS which included simple
multi-tasking supervisor plus high level VMCF support (high enuf so
implementation could be VMCF or IUCV or ... ). Allows you to write all
sorts of service virtual machine functions at very high level (even in
PLI) and be able to run it multi-task. Code amounts to 20-50 lines of
modification to existing CMS nucleus plus new resident module or two
(total lines somewhere between 500 and 1500). Multi-tasking is general
enuf that can be used even if you aren't using VMCF functions.
... snip ... top of post, old email index
and:
Date: 12/22/80 17:54
To: wheeler
Just send you a taskforce report covering CMS for the 811
hardware. Please do not discuss it with anyone outside Research
division or CSC until after we take it to Pok. (who asked for it). We
plan to go up on the 23 of December.
Please send me any comments you might have.
... snip ... top of post, old email index
past posts in this thread:
https://www.garlic.com/~lynn/2011b.html#19 A brief history of CMS/XA, part 1
https://www.garlic.com/~lynn/2011b.html#20 A brief history of CMS/XA, part 1
https://www.garlic.com/~lynn/2011b.html#23 A brief history of CMS/XA, part 1
https://www.garlic.com/~lynn/2011b.html#29 A brief history of CMS/XA, part 1
https://www.garlic.com/~lynn/2011b.html#33 A brief history of CMS/XA, part 1
https://www.garlic.com/~lynn/2011b.html#47 A brief history of CMS/XA, part 1
https://www.garlic.com/~lynn/2011b.html#52 A brief history of CMS/XA, part 1
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: Dyadic vs AP: Was "CPU utilization/forecasting" Newsgroups: bit.listserv.ibm-main Date: Wed, 20 Apr 2011 10:04:53 -0400martin_packer@UK.IBM.COM (Martin Packer) writes:
370AP ... was a two processor configuration where only one of the processors had channels ... the second processor purely did compute bound work. it was less expensive and could be applicable to more compute intensive work. it was also applicable in large loosely-coupled environment when running out of bcontrollerb interfaces for all the channels ... aka with four channel dasd controller with string-switch (for each disk connected to two controllers) ... giving 8 channel paths to each disk ... it would be possible to have eight two-processor complexes (for 16 processors total).
DYADIC was term introduced with 3081 ... where it wasn't possible to split the configuration and run as two separate independent processors (wanted to draw distinction between past 360/370 multiprocessor that could be split and run independently and the 3081 which couldn't be split). 3081 had both processors being able to address all channels and also introduced 31-bit virtual addressing.
trivia ... 360/67 had 32-bit virtual addressing, all processors could address all channels *AND* configuration could be split into independent running single processors. 360/67 was desgined for four-processor configuration, but I know of only a couple three-processor configuration that were actually built (and no four processor configurations) ... all the rest multiprocessor configurations were simply two-processors.
other 3081 trivia ... 370 (& 3081) dual-processor slowed machine cycle down by ten percent to help with multiprocessor processor cache interaction ... so a two-processor machine started out at only 1.8 times a single processor machine. Multiprocessor software and actual multiprocessor cache interactions tended to add additional overhead so that dual-processor tended to have 1.4-1.5 times the throughput of single processor.
3081 originally never intended to have single processor version ... but largely because ACP/TPF didn't have multi-processor support, there was eventually a 3083 introduced. The easiest would have been to remove 2nd processor from the 3081 box ... however, processor0 was at the top of the box and the 2nd processor1 was in the middle of the box ... which would have left the box dangerously top heavy.
eventually 3083 was introduced with single processor ... it was possible to turn-off the ten percent machine cycle slowdown (done for multiprocessor cache interaction) ... and eventually there was a special microcode load tuned for the ACP/TPF workloads that tended to be more I/O intensive.
in the late 70s, the consolidated internal online US HONE operation (US
HONE and the various HONE clones providied online world-wide sales &
marketing support) was the largest single-system operation in the world.
It was large loosely-coupled operation with "AP" multiprocessors ...
most of the sales&marketing applications were implemented in APL and the
workload was extremely compute intensive. I provided them with their
initial multiprocessor support ... highly optimized kernel
multiprocessor pathlengths and games played to improve cache hit
locality ... could get slightly better than twice single processor (i.e.
cache games offset the machine running at only 1.8times single processor
and the optimized multiprocessor pathlengths). misc. past posts
mentioning multiprocessor support (&/or compare&swap instruction)
https://www.garlic.com/~lynn/subtopic.html#smp
as mentioined in previous post
https://www.garlic.com/~lynn/2011f.html#41 CPU utilization/forecasting
the science center had done a lot of the early work in performance
monitoring, reporting, simulation, modeling, workload&configuration
profiling ... that evolves into capacity planning. misc. past posts
mentioning science center
https://www.garlic.com/~lynn/subtopic.html#545tech
One of the APL models was packaged in the mid-70s as the performance
predictor on HONE ... so that sales&marketing could take customer
workload&configuration specification and ask "what-if" questions about
workload &/or configuration changes. another version of the "model" was
modified and used to decide (online) workload balancing across the
loosely-coupled configuration (which processor complex would new logon
be directed to, part of single-system operation).
misc. past posts mentioning HONE
https://www.garlic.com/~lynn/subtopic.html#hone
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: Dyadic vs AP: Was "CPU utilization/forecasting" Newsgroups: bit.listserv.ibm-main Date: 20 Apr 2011 12:09:13 -0700lynn@GARLIC.COM (Anne & Lynn Wheeler) writes:
during the FS period ... there was lots of internal politics going on
and 370 product efforts were shutdown (as being possibly competitive).
When FS died, there was all sort of mad rush to get things back into the
370 product pipelines (the distraction of FS is claimed to have also
allowed clone processors to gain market foothold). ... this discusses
some of the FS effort, the internal politics and the dark shadow that
the FS failure cast over the corporation for decades
https://people.computing.clemson.edu/~mark/fs.html
part of the mad rush was to get out q&d 303x. The integrated channel microcode from the 158 was split off into 303x "channel director" (158 engine with only integrated channel microcode and no 370 microcode). A 3031 was a 158 engine (& new panels) with just the 370 microcode (and no integrated channel microcode) and a 2nd 158 engine (channel director) with just the integrated channel microcode. A 3032 was a 168-3 with new panels and 303x channel director(s) aka 158 engine. A 3033 started out being 168-3 logic remapped to 20% faster chips and 303x channel director(s) ... before the 3033 shipped there was some redoing logic (the chips were 20% faster but also had ten times as much circuits per chip ... initially went unused) ... which got 3033 up to about 50% faster than 168-3.
In parallel with 303x ... there was work on "811" architecture
(supposedly named for nov78 date on lots of the documents) and the 3081
(a FS machine with just 370 emulation microcode). Some of this is
discussed in this internal memo reference from the period:
http://www.jfsowa.com/computer/memo125.htm
A number of 3033up was starting to feel significantly memory/storage constrained being limited to maximum of 16mbyte real storage. It was much worse for 3033mp since they were also limited to same 16mbyte real storage constriant.
A hack was done for 3033 to allow more real storage ... even tho instruction addressing was limited to 24bit/16mbyte addressing. The 370 page table entry had two unused bits ... and the 3033 hack was to re-assign the two unused bits to prepend them to the page number, allowing specifying up to 2**14 4kbyte pages (aka 26bit/64mbyte). Real & virtual instruction addressing was still limited to 16mbytes ... but a page table entry could translate that up to 64mbytes. I/O then was done with IDALs ... which already had 31bit field.
MVS was (also) having severe problems at larger 3033 installations. Transition from OS/VS2 svs to OS/VS2 mvs ... involved mapping 8mbyte MVS kernel image into every 16mbyte application virtual address space. In order to support subsystems that were now (also) in separate virtual address working with application address space ... there was "common segment" in every virtual address space, that started out as 1mbyte (applications being able to place parameters in CSA and use pointer-passing API in subsystem call). For large 3033 MVS instalaltions, CSAs were hitting 4-5mbytes and threatening to grow to 5-6mbytes (leaving only 2mbytes for application). An internal shop that was a large multiple machine MVS operation had 7mbyte fortran chip design application. The MVS operation was carefully configured to keep CSA to 1mbyte ... and constant ongoing activity keeping the chip design Fortran application from exceeding 7mbyte. They were being faced with having to convert all their machines to vm/cms ... since CMS could allow the application to have nearly all of the 16mbyte virtual address space.
Early 3081Ds were shipped w/o the "811" extensions (vanilla 370 w/o 370-xa) and were supposedly slightly faster than 3033 ... however, there was a number of benchmarks that had 3081Ds 20% slower than 3033. 3081K came out with double the processor cache size ... supposedly 50% faster than 3033 ... but some benchmarks coming in only 5% faster than 3033.
The internal memo (mentioned above) goes into the enormous amount circuits-hardware in 3081 compared to its performance (especially when stacked up against clone processors).
tieing two 3081s into a 4-way 3084 really drove up the multiprocessor cache interference (invalidate signals coming from three other caches rather than one other cache). Both VM/HPO and MVS had kernel "cache sensitivity" work that involved carefully aligning kernel control areas & storage allocation on cache line boundaries (and making them multiple of cache lines). the issue was to prevent two different storage areas (which might be concurrently used by different processors) from occupying different parts of the same cache line (resulting in cache line thrashing being moved repeatedly between different caches). This change supposedly gained greater than 5% aggregate system throughput.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: US HONE Datacenter consolidation Newsgroups: alt.folklore.computers Date: Thu, 21 Apr 2011 09:14:01 -0400The US HONE datacenters were consolidated in Palo Alto in the mid-70s. At the time, it was open land next door. Now the land has bldg. where the president visited yesterday. The old US HONE datacenter bldg. now has a different occupant.
HONE had originated in the wake of the 23Jun69 unbundling announcements to give the young system engineers in branch offices a way of getting their "hands-on" operating practice (in virtual machines). Originally it was some number of 360/67 datacenters running (virtual machine) cp67.
It morphed into providing online APL-based applications supporting sales&marketing, first with cp67 cms\apl ... then transitioning to vm370 apl\cms (and the se "hands-on" activity dwindled away to nothing).
misc. past posts mentioning HONE
https://www.garlic.com/~lynn/subtopic.html#hone
misc. past posts mentioning 23jun69 unbundling
https://www.garlic.com/~lynn/submain.html#unbundle
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 23 Apr, 2011 Subject: Are Americans serious about dealing with money laundering and the drug cartels? Blog: Financial Crime Risk, Fraud and SecurityHow a big US bank laundered billions from Mexico's murderous drug gangs
from above:
As the violence spread, billions of dollars of cartel cash began to
seep into the global financial system. But a special investigation by
the Observer reveals how the increasingly frantic warnings
... snip ...
post from early jan. in (financial crime risk, fraud and security)
"What do you think about fraud prevention in governments?"
https://www.garlic.com/~lynn/2011.html#50 What do you think about fraud prevention in the governments?
part of the reference was that with the gov. bending over backwards to do everything possible to keep the too-big-to-fail institutions in business ... that with a little thing like drug cartel money laundering ... the gov. has been asking them to please stop.
Too Big to Jail - How Big Banks Are Turning Mexico Into Colombia
http://www.taipanpublishinggroup.com/tpg/taipan-daily/taipan-daily-080410.html
Banks Financing Mexico Gangs Admitted in Wells Fargo Deal
http://www.bloomberg.com/news/2010-06-29/banks-financing-mexico-s-drug-cartels-admitted-in-wells-fargo-s-u-s-deal.html
Wall Street Is Laundering Drug Money And Getting Away With It
http://www.huffingtonpost.com/zach-carter/megabanks-are-laundering_b_645885.html?show_comment_id=53702542
Banks Financing Mexico Drug Gangs Admitted in Wells Fargo Deal
http://www.sfgate.com/cgi-bin/article.cgi?f=/g/a/2010/06/28/bloomberg1376-L4QPS90UQVI901-6UNA840IM91QJGPBLBFL79TRP1.DTL
How banks aided drug traffic
http://www.charlotteobserver.com/2010/07/04/1542567/how-banks-aided-drug-traffic.html
The Banksters Laundered Mexican Cartel Drug Money
http://www.economicpopulist.org/content/banksters-laundered-mexican-cartel-drug-money
Money Laundering and the Global Drug Trade are Fueled by the Capitalist Elites
http://www.globalresearch.ca/index.php?context=va&aid=20210
Wall Street Is Laundering Drug Money and Getting Away with It
http://www.alternet.org/economy/147564/wall_street_is_laundering_drug_money_and_getting_away_with_it/
Money Laundering and the Global Drug Trade are Fueled by the Capitalist Elites
http://dandelionsalad.wordpress.com/2010/07/23/money-laundering-and-the-global-drug-trade-are-fueled-by-the-capitalist-elites-by-tom-burghardt/
Global Illicit Drugs Trade and the Financial Elite
http://www.pacificfreepress.com/news/1/6650-global-illicit-drugs-trade-and-the-financial-elite.html
Wall Street Is Laundering Drug Money And Getting Away With It
http://institute.ourfuture.org/blog-entry/2010072814/megabanks-are-laundering-drug-money-and-getting-away-it
Money laundering does carry a bit heavier fine ... including shutting down the financial institution and putting the executives in jail. However, when its the too-big-to-fail institutions are involved ... they seem reluctant to impose such measures.
There have been claims that china's measures (after throwing off the British yoke of imposed drug selling) are (in the long run) a lot more humane (and effective) method of dealing with the problem.
A counter scenario (at least with respect to SEC) was the congressional testimony by the person that tried for a decade to get the SEC to do something about Madoff ... or the GAO reports reviewing public company financial filing showing big uptick in fraudulent filings ... even after SOX (which theoretically had significant increased penalties, including executives going to jail) ... recent quote off the web: Enron was a dry run and it worked so well it has become institutionalized.
Corrupt Bank Oversight Is Creating New Immoral Hazard
http://blogs.forbes.com/neilweinberg/2011/04/14/corrupt-bank-oversight-is-creating-new-immoral-hazard/
In Financial Crisis, No Prosecutions of Top Figures
http://www.nytimes.com/2011/04/14/business/14prosecute.html?_r=2&hp
related to "corrupt bank oversight":
The Real Housewives of Wall Street; Why is the Federal Reserve forking
over $220 million in bailout money to the wives of two Morgan Stanley
bigwigs?
http://www.rollingstone.com/politics/news/the-real-housewives-of-wall-street-look-whos-cashing-in-on-the-bailout-20110411?page=1
In the case of GAO review of public company financial filings ... and the testimony of the person that tried for a decade to get the SEC to do something about Madoff ... it was lack of regulation enforcement. The person that had tried for a decade to get SEC to do something about Madoff, also mentioned something about SEC was all lawyers and nearly nobody with forensic accounting experience (this was also somewhat Cramer's comment about there was little risk in illegal naked short selling because wasn't anybody at the SEC that understood). In one of the recent news references upstream ... it may also have been pressure by members of congress (who possibly would have been implicated).
However, there is also some amount of deregulation (by congress) of some number of things.
Phil Gramm's Enron Favor
https://web.archive.org/web/20080711114839/http://www.villagevoice.com/2002-01-15/news/phil-gramm-s-enron-favor/
from above:
A few days after she got the ball rolling on the exemption, Wendy
Gramm resigned from the commission. Enron soon appointed her to its
board of directors, where she served on the audit committee, which
oversees the inner financial workings of the corporation. For this,
the company paid her between $915,000 and $1.85 million in stocks and
dividends, as much as $50,000 in annual salary, and $176,000 in
attendance fees,
... snip ...
People to Blame for the Financial Crisis; Phil Gramm
http://content.time.com/time/specials/packages/article/0,28804,1877351_1877350_1877330,00.html
from above:
He played a leading role in writing and pushing through Congress the
1999 repeal of the Depression-era Glass-Steagall Act, which separated
commercial banks from Wall Street. He also inserted a key provision
into the 2000 Commodity Futures Modernization Act that exempted
over-the-counter derivatives like credit-default swaps from regulation
by the Commodity Futures Trading Commission. Credit-default swaps took
down AIG, which has cost the U.S. $150 billion thus far.
... snip ...
Gramm and the 'Enron Loophole'
http://www.nytimes.com/2008/11/17/business/17grammside.html
from above:
Enron was a major contributor to Mr. Gramm's political campaigns, and
Mr. Gramm's wife, Wendy, served on the Enron board, which she joined
after stepping down as chairwoman of the Commodity Futures Trading
Commission.
... snip ...
Greenspan Slept as Off-Books Debt Escaped Scrutiny
http://www.bloomberg.com/apps/news?pid=20601109&refer=home&sid=aYJZOB_gZi0I
from above:
That same year Greenspan, Treasury Secretary Robert Rubin and SEC
Chairman Arthur Levitt opposed an attempt by Brooksley Born, head of
the Commodity Futures Trading Commission, to study regulating
over-the-counter derivatives. In 2000, Congress passed a law keeping
them unregulated.
... snip ...
Born must have been fairly quickly replaced by Gramm's wife, before she then left to join Enron (and the Enron audit committee) Gramm's wife apparently was put in as Born's replacement as a temporary stop-gap until Gramm got law passed that exempted regulation.
recent quote seen on the web: Enron was a dry run and it worked so well it has become institutionalized
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: The first personal computer (PC) Newsgroups: alt.folklore.computers Date: Sat, 23 Apr 2011 10:00:26 -0400hda <agent33@xs4all.nl_invalid> writes:
nominal mainframe 327x terminals were half-duplex and had annoying habit of locking the keyboard if you attempted to press a key at the same time the system was trying to write to the screen (which then required stopping and hitting reset to unlock the keyboard).
there was a small fifo box built ... unplug the keyboard from the head plug the fifo box into the head, and then plug the keyboard into the fifo box ... and the problem was eliminated.
the repeat function (like holding cursor movement down) was also extremely slow ... both the delay before starting repeat and the repeat rate. open up the keyboard and solder a couple resisters ... choice of value would select delay interval before starting repeat and repeat rate.
the only problem was that .1/.1 sec. value ... would get ahead of the screen update ... using cursor positioning ... say holding down cursor forward key ... the cursor would appear to coast for some distance after releasing. It took a little bit of practice to get use the timing of releasing held key ... and have the cursor stop at the correct positon.
then for the next generaiton 3274/3278 ... they moved quite a bit of the electronics from the head back into the 3274 controller ... eliminating any ability to do (human factor) hacks on the terminal. complaining to the plant ... and they eventually came back that the 3278 wasn't designed for interactive computing ... it was designed for data entry (aka fancy keypunch).
moving all the electronics back to the controller drove up the chatter
over the coax between controller & terminal ... as well as latency ...
terminal response for 3278 was significantly slower than 3277 (over and
above any system response). later with IBM/PC and terminal emulation
... upload/download rate for emuulated 3277 was three times faster than
emulated 3278 (because of difference in coax chatter and latency). misc.
past posts mentioning terminal emulation
https://www.garlic.com/~lynn/subnetwork.html#terminal
old post with some (real) 3277/3278 comparisons
https://www.garlic.com/~lynn/2001m.html#19 3270 protocol
and whether or not it was possible to achieve .1 second response.
recent mainframe related postings on the subject:
https://www.garlic.com/~lynn/2011d.html#53 3270 Terminal
https://www.garlic.com/~lynn/2011e.html#32 SNA/VTAM Misinformation
https://www.garlic.com/~lynn/2011e.html#94 coax (3174) throughput
https://www.garlic.com/~lynn/2011f.html#0 coax (3174) throughput
https://www.garlic.com/~lynn/2011f.html#33 At least two decades back, some gurus predicted that mainframes would disappear
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 23 Apr, 2011 Subject: At least two decades back, some gurus predicted that mainframes would disappear Blog: Greater IBMre:
... but so many died off that they've come close to becoming endangered species. at the height of the genocide there was internal references would the last person to leave POK, please turn off the lights (20yrs ago was when the company went into the red)
It was only a few years prior to going into the red ... that top executives were saying the "booming" mainframe business was going to double company revenue from $60B/yr to $120B/yr ... and they had a huge bldg. program to double (mainframe) manufacturing capacity ... however, the handwriting was already on the wall (disparity between the massive internal bldg program and the direction things were heading)
also see upthread reference to senior disk engineer at internal communication group world-wide annual conference (towards the end of the manufacturing "doubling" bldg program) and opening the talk with the statement that the communication group was going to be responsible for the demise of the disk division (because of the stranglehold that the communication group had on the datacenter). this happened just before the company was heading into the red
past posts in thread:
https://www.garlic.com/~lynn/2011e.html#15 At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened
https://www.garlic.com/~lynn/2011e.html#16 At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened
https://www.garlic.com/~lynn/2011e.html#17 At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened
https://www.garlic.com/~lynn/2011e.html#19 At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened
https://www.garlic.com/~lynn/2011f.html#32 At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened
https://www.garlic.com/~lynn/2011f.html#33 At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened
https://www.garlic.com/~lynn/2011f.html#35 At least two decades back, some gurus predicted that mainframes would disappear
https://www.garlic.com/~lynn/2011f.html#37 At least two decades back, some gurus predicted that mainframes would disappear
https://www.garlic.com/~lynn/2011f.html#39 At least two decades back, some gurus predicted that mainframes would disappear
https://www.garlic.com/~lynn/2011f.html#42 At least two decades back, some gurus predicted that mainframes would disappear
https://www.garlic.com/~lynn/2011f.html#45 At least two decades back, some gurus predicted that mainframes would disappear
https://www.garlic.com/~lynn/2011f.html#46 At least two decades back, some gurus predicted that mainframes would disappear
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 23 Apr, 2011 Subject: Are Americans serious about dealing with money laundering and the drug cartels? Blog: Financial Crime Risk, Fraud and Securityre:
double check who became ceo&chairman (of Citi) and brought in a lot of his people ... independent of what the company was called ... something similar happened to BofA (jokes about neither original Citi nor original BofA now exist).
The rhetoric on the floor of congress regarding GLBA was that its primary purpose was that institutions that were already banks got to remain banks ... but if you weren't already a bank, you didn't get to become a bank (aka bank modernization act) ... specifically calling out walmart and m'soft.
walmart then wanted to get one of the existing ILC charters ... similar to what amex got and a number of car companies (for making car loans). walmart's purpose was to become its own "acquiring" institution (to save that part of card acquiring transaction fees). Note that walmart accounts for 25-30% of all retail store transactions ... which would have been a big hit to walmart's acquiring institution (one of the too-big-to-fail operations) ... they managed to rally the community banks to lobby congress against walmart obtaining ILC (somehow that major hit to revenue of a too-big-to-fail institution would impact community banks).
roll forward to the bailouts ... majority coming from federal reserve behind the scenes (some 20+k pages forced to recently release by court order) ... but only for depository institutions with bank charters. A few of the wall street too-big-to-fail institutions didn't have bank charters already ... so federal reserve just gave them bank charters (which should have been in violation of glba).
we had been tangentially involved in the cal. state data breach notification legislation ... having been brought in to help wordsmith the cal. state electronic signature legislation. several of the participants in electronic signature were also heavily into privacy and consumer issues. they had done large, in-depth consumer surveys and found #1 issue was identity theft ... namely account fraud because of data breaches. At the time little or nothing was being done about the breaches (the resulting fraud was against the consumers and not the institutions having the breaches). There was some hope that the resulting publicity from the breach notifications would motivate institutions to start doing something about the fraud/breaches.
these organizations were also in the process of doing a cal. state personal information "opt-in" legislation (i.e. institutions can only use and share personal information when the person explicitly authorizes it). A federal pre-emption provision was also thrown into GLBA for "opt-out" ... i.e. institutions can use and share personal information as long as the person doesn't call up and complain.
A few years ago, I was at a national annual privacy conference in WASH DC which had a panel discussions with the FTC commissioners. somebody got up in the back of the room and asked if any of them were going to do anything about GLBA "opt-out". He said he was involved in operations of call centers used by many of the financial institutions and "knew" that the people answering the "opt-out" phone lines weren't provided any means of recording who called (and wanted to "opt-out" of privacy sharing, aka "opt-out" you need recorded proof that you objected ... however, in "opt-in" they need recorded proof you agreed). The "opt-out/out-in" issue has recently also come up with regard to some of the social networking sites (an instance of both regulatory "change" as well as failing to enforce regulation)
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: Drum Memory with small Core Memory? Newsgroups: alt.folklore.computers, comp.arch Date: Sat, 23 Apr 2011 15:01:46 -0400"Dave Wade" <dave.g4ugm@gmail.com> writes:
science center ... had started out by doing hardware modifications to 360/40 to support virtual memory (they had originally tried to get 360/50 ... but they were in scarce supply because so many were going to FAA for ATC system) ... and did (virtual machine) cp40. when 360/67 became available, they replace the 360/40 and cp40 morphed into cp67.
360/67 came with 2301 "paging drum" ... 4mbyte capacity. it was similar to the 2303 drum ... but read/wrote four heads in parallel ... so had four times the transfer rate.
three people came out to univ in jan68 to install cp67 (our 360/67 ran a little tss/360 on weekends ... but rest of the time operated with os/360 ... not making use of virtual memory). they used tss/360 drum format ... which had nine 4kbyte pages across two tracks (one of the pages spanned the end of one track and the start of the next). however, the i/o queue was managed FIFO with single request at a time ... 2301 peaked at 80 page transfers/sec. I redid the i/o stuff ... ordered seek queuing (for moveable arm disk) and ordered chained requests (chaining queued requests on same position ... applicable to queued req for same disk cyl and/or whole drum). The change resulted in nearly doubling effective disk active and "graceful degradation" as load increased ... and allowed 2301 to peak at nearly 300 4k transfers/sec.
univ. originally had 512kbyte "core" 360/67 ... but it got upgraded to 768kbyte 360/67 fairly early because tss/360 was becoming so bloated (and then they still didn't use tss/360).
there was tss/360 benchmark involving single processor 360/67 with 1mbyte memory and a two-processor 360/67 with 2mbyte of memory where the 2-processor version had 3.8 times the throughput as the 1-processor. Both benchmarks were still quite dismall ... but somebody in the TSS camp tried to hype pixie dust that tss/360 could magically increase thruput (nearly four times the thruput with only twice the resources). The actualy situation was that tss/360 kernel real storage requirements were so bloated that in 1mbyte of real memory ,,, there was very little real memory left for applications. In 2-processor, 2mbyte configuration, kernel real storage didn't double ... so there was possibly four times the real storage left for applications resulting in 3.8 times the thruput (because of significantly reduction in real storage contention and page thrashing).
The SE at the univ. responsible for tss/360 and I did a benchmark script ... fortran source, input/edit, compile and execute the same program (with emulated "think" & human time). I did it for cms ... and with 35 simulated cms users running the script got better response and thruput than he did running the same script with just four TSS/360 users (same hardware and configuration)
misc. recent posts mentioning 2301 (and/or its follow-on, the 2305 fixed
head disk):
https://www.garlic.com/~lynn/2011b.html#69 Boeing Plant 2 ... End of an Era
https://www.garlic.com/~lynn/2011b.html#72 IBM Future System
https://www.garlic.com/~lynn/2011e.html#1 Multiple Virtual Memory
https://www.garlic.com/~lynn/2011e.html#54 Downloading PoOps?
https://www.garlic.com/~lynn/2011e.html#75 I'd forgotten what a 2305 looked like
https://www.garlic.com/~lynn/2011e.html#79 I'd forgotten what a 2305 looked like
https://www.garlic.com/~lynn/2011f.html#46 At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: Are Tablets a Passing Fad? Newsgroups: alt.folklore.computers Date: Sun, 24 Apr 2011 09:29:07 -0400Morten Reistad <first@last.name> writes:
m'soft revenue was pegged to everybody always buying the new version (and some growth because worldwide market hadn't completely saturated) ... if that changes to every other new version ... revenue could be cut in half ... but it doesn't necessarily/yet mean that majority of the people have stopped using m'soft software.
in the late 90s, it was motivation to play with fee per use revenue (somewhat like oldtime mainframe leasing model or oldtime commercial timesharing) ... which has now being associated with "cloud computing".
some of the tablet hype ... is with relatively small market penetration ... any growth is seen as significant growth ... and some of the folks that are constantly buying new toys ... have something else to buy (besides or instead of the latest PC). Lots of hype is 1st & 2nd order (change) in sales numbers.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: Are Tablets a Passing Fad? Newsgroups: alt.folklore.computers Date: Sun, 24 Apr 2011 11:43:45 -0400Stan Barr <plan.b@dsl.pipex.com> writes:
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: Drum Memory with small Core Memory? Newsgroups: alt.folklore.computers, comp.arch Date: Mon, 25 Apr 2011 16:41:11 -0400John Levine <johnl@iecc.com> writes:
transfer rate being approx. equal ... then fixed head drum/disk makes a lot more difference for demand page operations ... since there was access latency for possibly large number of operations. if the swap transfers were set up appropriately then there was just a single access latency for lengthy transfer.
on mainframe going from 3330 to 3380 ... there was only modest improvement in access latency but a factor of ten times change in transfer rate. "big pages" were created in MVS & VM to leverage the change ... where on outbound ... a full track of (10) pages were collected for the write ... and a page fault for any of the ten pages would transfer all ten pages in one operation.
big pages tended to increase real storage requirements and inflate working sets ... at a trade-off of drastically reduced number of separate access (sort of part way between "small" demand paging and all out swapping).
the issue was that the early implementations tended to totally trash the page replacement accuracy ... degrading the trade-off benefits.
old email about trying to correct a whole bunch of implementations
problems in "big page" implementation (that had existed for several
years)
https://www.garlic.com/~lynn/2011e.html#email870320
in this post
https://www.garlic.com/~lynn/2011e.html#27 Multiple Virtual Memory
the above email mentions the term "SWAP" ... which actually is referring to "big page" implementation. the reference to "STC drum" is "electronic" device simulating fixed-head device with no rotational latency (where defined for "big pages", system ran at 70% utilization, but running w/o "big pages", system ran at 100% utilization ... effectively all higher thruput ... since paging operations had very small pathlength). above also has long list of past posts mentioning "big pages"
past posts mentioning page replacement work
https://www.garlic.com/~lynn/subtopic.html#wsclock
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: Dyadic vs AP: Was "CPU utilization/forecasting" Newsgroups: bit.listserv.ibm-main Date: 25 Apr 2011 14:50:31 -0700In <005a01cbff6c$b63dded0$22b99c70$@hawkins1960@sbcglobal.net>, on
re:
https://www.garlic.com/~lynn/2011f.html#49 Dyadic vs AP: Was "CPU utilization/forecasting"
https://www.garlic.com/~lynn/2011f.html#50 Dyadic vs AP: Was "CPU utilization/forecasting"
360&370 had two-processor multiprocessor shared memory and although had dedicated channels ... tended to try and simulate shared channels by trying to configure the same channel numbers on the two processors so they connected to same controllers (at same addresses, for controllers that supported multiple channel attachments) ... allowing I/O to be done to the same controller/device by both processors.
370-APs only had one of the processors with channels ... the other processor was purely for dedicated execution. I/O requests that originated on the attahced processor (w/o channels ... or in multiprocessor for device only connected to channel on the other processor) ... resulted in internal kernel operation that handed off the i/o request to processor with the appropriately connected channel.
one of the issues in cache machines ... was that high interrupt rates tended to have very deterious effect on cache-hit ratios (translates to effective MIP rate) ... where cache entries for the running application got replaced with cache entries for interrupt & device i/o handling ... and then possibly replaced again when switching back to the running application.
A gimick I hased in the early/mid 70s for cache machine ... was when observed I/O interrupt rates exceeding threshold ... started running disabled for I/O interrupts ... but with periodic timer interrupt. At the timer interrupt, all pending interrupts would be "drained" under software control (using SSM to enable for i/o interrupts). The increased delay in taking the i/o interrupt was more than offset by both increased cache hit ratio (aka MIP rate) of the application running w/o interrupts, and then increased cache hit ratio (aka MIP rate) of effectively batch processing multiple I/O interrupts.
For AP support, I also had a gimick that tended to keep the CPU intensive operations on the processor w/o channels (sort of natural CPU affinity). two processor 370 cache machine operation slowed down the processor machine cycle by 10% (for multiprocessor operation to take into account cross-cache communication) ... resulted in two-processor having base hardware of 1.8 times single processor ... multiprocessor software overhead then tended to result in multiprocssor having 1.4-1.5 times that of uniprocessor.
For HONE 370APs, sometimes I could get 2-processor throughput more than
twice single processor thruput. HONE was heavily compute intensive APL
applications ... although some would periodically do lots of I/O. The
natural processor/cache affinity (improved MIP rate) increased thruput
(along with having extremely short multiprocessor support pathlengths)
... keeping the compute intensive (non I/O) application execution on the
AP processor w/o channels. misc. past posts mentioning (virtual machine
based) HONE (US HONE & HONE clones around the world provided world-wide
sales & marketing support)
https://www.garlic.com/~lynn/subtopic.html#hone
This got messed up in the early 3081 dyadic time-frame. Since ACP/TPF didn't have multiprocessor support (and it hadn't yet been decided to do 3083) ... VM was "enhanced" to try and improve ACP/TFP 3081 virtual machine thruput. For a lot of VM pathlength that used to be serialized with virtual machine execution ... there was an attempt to make it asynchronous ... running on the 2nd, presumably idle processor ... with lots of request queuing and processor "shoulder taping" (the increase in overhead theoretically offset by the reduction in ACP/TPF elapsed time). However, for customers that had been running fully loaded (non-ACP/TPF) multiprocessor operation ... the transition to this new release represented significant degradation (the increased request queuing and shoulder taping taking 10-15% of both processors).
Then there was a number of onsite visits at various large customers ... attempting to perform other kinds of tuning operations to mask the enormously increased multiprocessor overhead in the new release (all the shoulder taping also messed up the natural affinity characteristics ... motivating a large increase in explicit specified processor affinity).
old email mentioning large gov. TLA ... trying to provide a variety of
performance improvements to offset the multiprocessor overhead increase:
https://www.garlic.com/~lynn/2001f.html#email830420
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: Drum Memory with small Core Memory? Newsgroups: alt.folklore.computers, comp.arch Date: Tue, 26 Apr 2011 09:21:55 -0400John Levine <johnl@iecc.com> writes:
2303 was about the same transfer as 2314 .... it was the 2301 fixed-head drum ... looked very much like 2303 and about same capacity ... that transferred four heads in parallel (and had four times the transfer rate).
earlier post in thread mentions more about 2301 as paging drum. it still had rotational latency ... and cp67 delivered to univ in jan1968 did FIFO queuing and single transfer per operation (peaked about 80 4kbyte page transfers per second). I redid for rotational order and multiple requests per i/o operation ... and could get nearly 300 4kbyte page transfers per second (over mbyte/sec).
2311 was 156kbytes/sec, 2314 was 310kbytes/sec
https://en.wikipedia.org/wiki/IBM_magnetic_disk_drives
recent thread about 2305 (fixed-head disk ... follow-on to 2301)
https://www.garlic.com/~lynn/2011e.html#75 I'd forgotten what a 2305 looked like
https://www.garlic.com/~lynn/2011e.html#79 I'd forgotten what a 2305 looked like
and possibility of 2305 attached to 360s (other than 360/85). this post
about spook base & igloo white .... mentions 2305 and 360/65s ... but
it is much more likely he met 2301 (2305 controller supported rotational
position sensing, and multiple exposures ... requiring 2880 block-mux
channel ... first available on 360/85):
https://web.archive.org/web/20030212092342/http://home.att.net/~c.jeppeson/igloo_white.html
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: Mixing Auth and Non-Auth Modules Newsgroups: bit.listserv.ibm-main Date: 26 Apr 2011 07:28:14 -0700jeff.holst@FISERV.COM (Jeff Holst) writes:
6670s (copier3 with computer interface added) were appearing in every
departmental area for distributed computer output. the 6670 driver had
been modified to include a randomly selected quote on the separator
pages. part of the audit was off-hours sweep of all the distributed
printers ... looking for sensitive output that was left out/unattended.
In one of the areas, the auditors found an output separator page with
the following quote:
[Business Maxims:] Signs, real and imagined, which belong on the walls
of the nation's offices:
1) Never Try to Teach a Pig to Sing; It Wastes Your Time and It Annoys the Pig.
2) Sometimes the Crowd IS Right.
3) Auditors Are the People Who Go in After the War Is Lost and Bayonet the Wounded.
4) To Err Is Human -- To Forgive Is Not Company Policy.
... snip ...
the next day, the auditors tried to escalate an issue that we were purposefully ridiculing them.
In the wake of Enron, congress passed sarbanes-oxley that significantly
increased audit requirements and penalties. A few years ago I was at a
financial conference in europe of european corporate CEOs and exchange
presidents ... major theme was that the (significant) SOX audit
requirements and costs were leaking out into the rest of the
world. There was semi-humorous reference to the country hosting the
conference on sunday cnn gps program
http://globalpublicsquare.blogs.cnn.com/2011/04/24/rent-the-country-of-liechtenstein-for-70k-a-night/
My position was that the increased audit requirements wouldn't make any significant dent in fraud (was more likely just a full-employment favor to the audit industry by congress) and possibly only significant part of SOX was section on informants. It turns out that apparently GAO also thot something similar and was doing reports of review of public company financial filings showing uptick in fraudulent filings after SOX (problem with both the audits and SEC enforcement).
In congressional testimony by the person that had tried for a decade to get SEC to do something about Madoff, there was mention that tips turn up 13 times more fraud than audits and SEC didn't have a tip hotline, but had a 1-800 line for companies to complain about audits.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: The IBM Selective Sequence Electronic Calculator Newsgroups: alt.folklore.computers Date: Tue, 26 Apr 2011 11:10:10 -0400Quadibloc <jsavard@ecn.ab.ca> writes:
recent referencs to cp67 ... virtual machine/virtual memory system
developed for 360/67 ...
https://www.garlic.com/~lynn/2011e.html#63 Collection of APL documents
https://www.garlic.com/~lynn/2011e.html#72 Collection of APL documents
https://www.garlic.com/~lynn/2011e.html#79 I'd forgotten what a 2305 looked like
https://www.garlic.com/~lynn/2011e.html#83 History of APL -- Software Preservation Group
https://www.garlic.com/~lynn/2011e.html#84 New job for mainframes: Cloud platform
https://www.garlic.com/~lynn/2011f.html#3 History of APL -- Software Preservation Group
https://www.garlic.com/~lynn/2011f.html#6 New job for mainframes: Cloud platform
https://www.garlic.com/~lynn/2011f.html#12 New job for mainframes: Cloud platform
https://www.garlic.com/~lynn/2011f.html#36 Early mainframe tcp/ip support (from ibm-main mailing list)
https://www.garlic.com/~lynn/2011f.html#39 At least two decades back, some gurus predicted that mainframes would disappear in future and it still has not happened
https://www.garlic.com/~lynn/2011f.html#51 US HONE Datacenter consolidation
https://www.garlic.com/~lynn/2011f.html#56 Drum Memory with small Core Memory?
https://www.garlic.com/~lynn/2011f.html#61 Drum Memory with small Core Memory?
done at the science center ... misc. past posts
https://www.garlic.com/~lynn/subtopic.html#545tech
cp67 installed at the univ. in jan1968 had terminal support for 1052 and 2741. the univ. had some number of tty/ascii terminals and I undertook to added tty/ascii support to cp67 ... consistent with the existing 1052&2741 support which leveraged the 2702 "SAD" command as part of doing dynamic terminal identification. I had hoped to have single "dial-in" number ("hunt-group") for all dialup terminals. My dynamic terminal-type support worked just fine for direct/leased lines ... but there was problem with dynamic dial-up lines. While it was possible to dynamically change line-scanner associated with each port/line using SAD command ... a short-cut in the 2702 hardwired the line-speed for each port.
this was somewhat motivation for the univ. to start clone controller
(that would be able to do both dynamic terminal type and dynamic line
speed) using an Interdata/3. The mainframe channel was reverse
engineered and channel interface board built for the Interdata/3 ...
programmed to simulate 2702 controller. Later four of us were written
up for being responsible for some part of clone controller business
... misc. past posts
https://www.garlic.com/~lynn/submain.html#360pcm
past posts mentioning MTS
https://www.garlic.com/~lynn/93.html#15 unit record & other controllers
https://www.garlic.com/~lynn/93.html#23 MTS & LLMPS?
https://www.garlic.com/~lynn/93.html#25 MTS & LLMPS?
https://www.garlic.com/~lynn/93.html#26 MTS & LLMPS?
https://www.garlic.com/~lynn/98.html#15 S/360 operating systems geneaology
https://www.garlic.com/~lynn/2000.html#89 Ux's good points.
https://www.garlic.com/~lynn/2000.html#91 Ux's good points.
https://www.garlic.com/~lynn/2000b.html#61 VM (not VMS or Virtual Machine, the IBM sort)
https://www.garlic.com/~lynn/2000c.html#44 WHAT IS A MAINFRAME???
https://www.garlic.com/~lynn/2000f.html#52 TSS ancient history, was X86 ultimate CISC? designs)
https://www.garlic.com/~lynn/2000g.html#0 TSS ancient history, was X86 ultimate CISC? designs)
https://www.garlic.com/~lynn/2001m.html#55 TSS/360
https://www.garlic.com/~lynn/2001n.html#45 Valid reference on lunar mission data being unreadable?
https://www.garlic.com/~lynn/2002i.html#63 Hercules and System/390 - do we need it?
https://www.garlic.com/~lynn/2002n.html#54 SHARE MVT Project anniversary
https://www.garlic.com/~lynn/2002n.html#58 IBM S/370-168, 195, and 3033
https://www.garlic.com/~lynn/2002n.html#64 PLX
https://www.garlic.com/~lynn/2003b.html#0 Disk drives as commodities. Was Re: Yamhill
https://www.garlic.com/~lynn/2003b.html#10 Disk drives as commodities. Was Re: Yamhill
https://www.garlic.com/~lynn/2003f.html#41 SLAC 370 Pascal compiler found
https://www.garlic.com/~lynn/2003k.html#5 What is timesharing, anyway?
https://www.garlic.com/~lynn/2003k.html#9 What is timesharing, anyway?
https://www.garlic.com/~lynn/2003l.html#30 Secure OS Thoughts
https://www.garlic.com/~lynn/2003l.html#41 Secure OS Thoughts
https://www.garlic.com/~lynn/2003m.html#32 SR 15,15 was: IEFBR14 Problems
https://www.garlic.com/~lynn/2004.html#46 DE-skilling was Re: ServerPak Install via QuickLoad Product
https://www.garlic.com/~lynn/2004.html#47 Mainframe not a good architecture for interactive workloads
https://www.garlic.com/~lynn/2004c.html#6 If the x86 ISA could be redone
https://www.garlic.com/~lynn/2004c.html#47 IBM 360 memory
https://www.garlic.com/~lynn/2004g.html#4 Infiniband - practicalities for small clusters
https://www.garlic.com/~lynn/2004l.html#16 Xah Lee's Unixism
https://www.garlic.com/~lynn/2004n.html#4 RISCs too close to hardware?
https://www.garlic.com/~lynn/2004n.html#25 Shipwrecks
https://www.garlic.com/~lynn/2004n.html#34 RISCs too close to hardware?
https://www.garlic.com/~lynn/2004o.html#20 RISCs too close to hardware?
https://www.garlic.com/~lynn/2004o.html#21 Integer types for 128-bit addressing
https://www.garlic.com/~lynn/2005.html#5 [Lit.] Buffer overruns
https://www.garlic.com/~lynn/2005.html#18 IBM, UNIVAC/SPERRY, BURROUGHS, and friends. Compare?
https://www.garlic.com/~lynn/2005g.html#56 Software for IBM 360/30
https://www.garlic.com/~lynn/2005k.html#20 IBM/Watson autobiography--thoughts on?
https://www.garlic.com/~lynn/2005s.html#17 winscape?
https://www.garlic.com/~lynn/2006c.html#18 Change in computers as a hobbiest
https://www.garlic.com/~lynn/2006d.html#6 IBM 610 workstation computer
https://www.garlic.com/~lynn/2006e.html#31 MCTS
https://www.garlic.com/~lynn/2006f.html#19 Over my head in a JES exit
https://www.garlic.com/~lynn/2006i.html#4 Mainframe vs. xSeries
https://www.garlic.com/~lynn/2006i.html#22 virtual memory
https://www.garlic.com/~lynn/2006k.html#41 PDP-1
https://www.garlic.com/~lynn/2006k.html#42 Arpa address
https://www.garlic.com/~lynn/2006m.html#42 Why Didn't The Cent Sign or the Exclamation Mark Print?
https://www.garlic.com/~lynn/2006m.html#47 The System/360 Model 20 Wasn't As Bad As All That
https://www.garlic.com/~lynn/2006n.html#43 MTS, Emacs, and... WYLBUR?
https://www.garlic.com/~lynn/2006o.html#3 MTS, Emacs, and... WYLBUR?
https://www.garlic.com/~lynn/2006o.html#36 Metroliner telephone article
https://www.garlic.com/~lynn/2007f.html#7 IBM S/360 series operating systems history
https://www.garlic.com/~lynn/2007f.html#62 What happened to the Teletype Corporation?
https://www.garlic.com/~lynn/2007j.html#6 MTS *FS tape format?
https://www.garlic.com/~lynn/2007m.html#60 Scholars needed to build a computer history bibliography
https://www.garlic.com/~lynn/2007q.html#15 The SLT Search LisT instruction - Maybe another one for the Wheelers
https://www.garlic.com/~lynn/2007t.html#54 new 40+ yr old, disruptive technology
https://www.garlic.com/~lynn/2007u.html#18 Folklore references to CP67 at Lincoln Labs
https://www.garlic.com/~lynn/2007u.html#23 T3 Sues IBM To Break its Mainframe Monopoly
https://www.garlic.com/~lynn/2007u.html#84 IBM Floating-point myths
https://www.garlic.com/~lynn/2007u.html#85 IBM Floating-point myths
https://www.garlic.com/~lynn/2007v.html#11 IBM mainframe history, was Floating-point myths
https://www.garlic.com/~lynn/2007v.html#32 MTS memories
https://www.garlic.com/~lynn/2007v.html#47 MTS memories
https://www.garlic.com/~lynn/2008d.html#94 The Economic Impact of Stimulating Broadband Nationally
https://www.garlic.com/~lynn/2008h.html#44 Two views of Microkernels (Re: Kernels
https://www.garlic.com/~lynn/2008h.html#78 Microsoft versus Digital Equipment Corporation
https://www.garlic.com/~lynn/2008q.html#65 APL
https://www.garlic.com/~lynn/2009j.html#77 More named/shared systems
https://www.garlic.com/~lynn/2009k.html#1 A Complete History Of Mainframe Computing
https://www.garlic.com/~lynn/2009k.html#70 An inComplete History Of Mainframe Computing
https://www.garlic.com/~lynn/2009l.html#34 August 7, 1944: today is the 65th Anniversary of the Birth of the Computer
https://www.garlic.com/~lynn/2009p.html#76 The 50th Anniversary of the Legendary IBM 1401
https://www.garlic.com/~lynn/2010g.html#34 someone smarter than Dave Cutler
https://www.garlic.com/~lynn/2010i.html#44 someone smarter than Dave Cutler
https://www.garlic.com/~lynn/2010j.html#37 Idiotic programming style edicts
https://www.garlic.com/~lynn/2010j.html#67 Article says mainframe most cost-efficient platform
https://www.garlic.com/~lynn/2010k.html#11 TSO region size
https://www.garlic.com/~lynn/2010p.html#42 Which non-IBM software products (from ISVs) have been most significant to the mainframe's success?
https://www.garlic.com/~lynn/2011.html#6 IBM 360 display and Stanford Big Iron
https://www.garlic.com/~lynn/2011.html#73 Speed of Old Hard Disks - adcons
https://www.garlic.com/~lynn/2011.html#86 Utility of find single set bit instruction?
https://www.garlic.com/~lynn/2011b.html#44 Colossal Cave Adventure in PL/I
https://www.garlic.com/~lynn/2011e.html#8 Multiple Virtual Memory
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 26 Apr, 2011 Subject: Are Americans serious about dealing with money laundering and the drug cartels? Blog: Financial Crime Risk, Fraud and Securityre:
others beside Gramm ... note treasury sec. ... formally of goldman
... helped with the citi deal and then went on to citi ... also served
stint as ceo of citi.
https://en.wikipedia.org/wiki/Citigroup
http://www.nytimes.com/2008/04/27/business/27rubin.html
http://www.npr.org/templates/story/story.php?storyId=15995005
http://in.reuters.com/article/2009/01/09/us-chronology-rubin-sb-idINTRE5086XS20090109
http://politifi.com/news/Why-Robert-Rubin-and-Citibank-Execs-Should-Be-in-Deep-Trouble-1330381.html
http://www.reuters.com/article/2009/01/09/chronology-rubin-idUSN0931351920090109?pageNumber=1
also mentioned above as being part of the group blocking Born
regarding commodity trading.
http://www.pbs.org/wgbh/pages/frontline/warning/
and while at citi tried to intervene on behalf of enron
http://www.truthdig.com/report/item/20080729_sucking_up_to_the_bankers/
bring it a little back to money laundering
Japan closes Citigroup branches
http://news.bbc.co.uk/2/hi/business/3666828.stm
Japan raps Citi for lax money laundering controls
http://www.reuters.com/article/2009/06/26/citigroup-japan-idUSN2528049820090626
Money laundering allegations hit Citibank Indonesia
http://www.asianewsnet.net/home/news.php?id=18570&sec=2
slightly humorous indirect money laundering from CNN GPS yesterday
http://globalpublicsquare.blogs.cnn.com/2011/04/24/rent-the-country-of-liechtenstein-for-70k-a-night/
the country had been on money laundering "black list" ... possibly the
swiss banks had been outsourcing the actual money laundering
transactions to the country next door.
http://www.indianexpress.com/Storyold/154519/
I was there a few years ago for a financial conference of European company and exchange CEOs (the conference theme focused on SOX audit substantial costs were leaking out into the rest of the world). The innkeeper made jokes about getting lots of people with business cards that read "dept. of money laundering" (implying that they were in charge of doing money laundering).
On the regulatory enforcement theme ... as mentioned upthread, GAO reports have shown that there was even uptick in public company fraudulent financial reports after SOX ... so other than full employment favor for audit companies, SOX appears to have little effect.
if one was cynical ... then regarding anti-money laundering outside the country ... there might be some comment about leveraging the gov. to eliminate competition
related to above about lobbying directed at preventing walmat (and others) cutting down on its interchange fees:
Swiped: Banks, Merchants And Why Washington Doesn't Work For You
http://www.huffingtonpost.com/2011/04/28/swipe-fees-interchange-banks-merchants_n_853574.html
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: The IBM Selective Sequence Electronic Calculator Newsgroups: alt.folklore.computers Date: Wed, 27 Apr 2011 08:37:20 -0400Charles Richmond <frizzle@tx.rr.com> writes:
then there is the opposite ... graduate student from some other dept ... was constantly coming by and demanding that somebody fix their programs so they ran correctly.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 27 Apr, 2011 Subject: Bank email archives thrown open in financial crash report Blog: Financial Crime Risk, Fraud and SecurityBank email archives thrown open in financial crash report
from above:
The internal email archives of ratings agencies and banks have been
thrown open as part of a major government investigation, demonstrating
the risk appetite of large Wall Street institutions before the global
economic crash.
... snip ...
A lot of this was in testimony in the fall2008 congressional hearings into the rating agencies
During the fall2008 congressional hearings into the rating agencies ... the comment was made that the rating agencies might blackmail the gov. into not prosecuting with the threat of downgrading the gov. credit rating .... the gov. credit rating has been in the news today.
There had been some securitized mortgages (CDOs) in the S&L crisis with doctored supporting documents for fraud. In the late 90s, we were asked to look at what could be done for integrity/assurance of securitized mortgages (CDOs) supporting documents.
In the first of this century, loan originators found they could pay rating agencies for triple-A ratings and immediately unload every loan (without regard to quality or borrower's qualifications). Speculators found that no-down, no-documentation, 1% interest payment only ARMs could make 2000% ROI buying&flipping properties in parts of the country with 20%-30% inflation.
Buyers of triple-A rated toxic CDOs didn't care about supporting documents, they were buying purely based on the triple-A rating. Rating agencies no longer cared about supporting documentation because they were being payed to give triple-A rating. Supporting documentation just slowed down loan originator's process of issuing loans. Since nobody cared about supporting documentations, they became superfluous ... which also resulted in there no longer being issue about supporting documentation integrity/assurance.
Lending money used to be about making profit on the loan payments over the life of the loan. With triple-A rated toxic CDOs, for many, it became purely a matter of the fees & commissions on the transactions and doing as many as possible. There were reportedly $27T in triple-A rated toxic CDO transactions done during the bubble ... with trillions in fees & commissions disappearing into various pockets.
Possibly aggregate 15%-20% take on the $27T ($5.4T) as the various transactions wander through the infrastructure (starting with original real estate sale). Reports were that wall street tripled in size (as percent of GDP) during the bubble. Also NY state comptroller had report that aggregate wall street bonuses spiked over 400% during the bubble.
and as been referenced in other discussions
http://blogs.forbes.com/neilweinberg/2011/04/14/corrupt-bank-oversight-is-creating-new-immoral-hazard/
and
http://www.nytimes.com/2011/04/14/business/14prosecute.html?_r=2&hp
there are various players with other motivations for not wanting to see prosecution (like many members of congress possibly having taken significant favors).
much of this came up in the fall2008 congressional hearings into the rating agencies (reference that rating agencies might blackmail the gov into not prosecuting with threat of credit downrating):
S&P's Credibility Under Fire As Agency Issues US Debt Warning;
Bipartisan Denate Report Cited S&P for Enabling US Mortgage Meltdown
http://abcnews.go.com/Politics/standard-poors-credibility-fire-us-debt-warning/story?id=13407823
Blog reference about lots of the news this month
http://www.winningprogressive.org/culprits-of-2008-financial-collapse-identified-world-ignores-5
has been to distract from this release:
"Wall Street and the Financial Crisis: Anatomy of a Financial Collapse"
http://levin.senate.gov/newsroom/supporting/2011/PSI_WallStreetCrisis_041311.pdf
Part of the overly prescriptive legislation may be show/obfuscation because they aren't addressing splitting apart the risky investment banking from safe&soundness of depository institutions (where the combination resulted in misaligned business processes and people motivated to do the wrong thing ... a side-effect of repeal of Glass-Steagall). Furthermore to be effective, legislation and regulation requires enforcement ... and while a lot was eliminated leading up to the financial mess ... there was still a whole lot that just weren't being enforced.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Old email from spring 1985 Newsgroups: alt.folklore.computers Date: Thu, 28 Apr 2011 09:50:30 -0400Date: 04/09/85 16:23:16
old reference to packaging large number of 370 & 801 chips in rack
https://www.garlic.com/~lynn/2004m.html#17 mainframe and microprocessor
more recent reference:
https://www.garlic.com/~lynn/2011.html#19 zLinux OR Linux on zEnterprise Blade Extension???
somewhat back-to-the-future with latest mainframe announcements.
http://www-03.ibm.com/systems/z/hardware/zenterprise/
Date: 04/09/85 16:31:14
From: wheeler
To: kingston
re: replies; didn't bother to reply since most of my comments about
global LRU and >16meg are detailed in VMPERF SCRIPT. Extra corswap
bits is easy fall-out of the 1.5bit algorithm change since it
scavenges the high-byte of CORPGPNT for (max) eight extra psuedo
reference bits already.
... snip ... top of post, old email index
past email mentioning global LRU
https://www.garlic.com/~lynn/lhwemail.html#globalru
Date: 04/09/85 17:44:01
From: wheeler
To: raleigh
fyi ... HSDT was presented to NSF ... and they are planning on using
it for backbone to connect all the super-computer centers. Over the
next 2-6 weeks will be presenting it to numerous other government &
research organizations.
... forgot, using SDLC chip for serialization/deserialization ... the
fastest one on the market only goes 2 megabits. HSDT won't use SDLC
because it will have to run at least 6 megabits ... and going to
higher rates later. Hardware people tell me that because of the
complexity of SDLC, chips for SDLC will continuelly lag most of the
various other alternatives ... either being much slower &/or more
expensive.
... snip ... top of post, old email index, NSFNET email
other old email mentioning hsdt
https://www.garlic.com/~lynn/lhwemail.html#hsdt
and nsfnet
https://www.garlic.com/~lynn/lhwemail.html#nsfnet
Date: 3 May 1985, 11:20:56 PST
To: wheeler
From: somebody at SLAC
I very much enjoyed the talk you gave on PAM in Toronto. I have done
some work at SLAC trying to make CMS do I/O faster, but PAM sounds
like a much better general solution. I would like to read the two
Research reports you mentioned. Could you tell me how I might get a
copy?
... snip ... top of post, old email index
misc. past posts mentioning paged mapped filesystem
https://www.garlic.com/~lynn/submain.html#mmap
post from earlier this year
https://www.garlic.com/~lynn/2011.html#4 Is email dead? What do you think?
with old email referencing using the low-level API to redo CP's spool
file implementation:
https://www.garlic.com/~lynn/2011.html#email870306
... and getting ready for the move up the hill (from bldg. 28) to
Almaden:
Date: 3 May 1985, 16:13:25 PDT
Subject: Friday
Busy week at 028 - Almaden planning swings into high gear, with the
computer lab being occupiable in less than 5 weeks. How about going
to Emperor Norton's and discussing some OTHER topics? (Or maybe just
eating and drinking ourselves into a frenzy?)
... snip ... top of post, old email index
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: program coding pads Newsgroups: alt.folklore.computers Date: Thu, 28 Apr 2011 11:03:16 -0400Lawrence Statton <lawrence@cluon.com> writes:
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: how to get a command result without writing it to a file Newsgroups: comp.lang.rexx Date: Thu, 28 Apr 2011 13:30:44 -0400"Dave Saville" <dave@invalid.invalid> writes:
I had earlier "paged" other storage associated with virtual memory infrastructure ... as had other virtual memory implementations ... especially single-level-store and paged-mapped architectures ... where some of these other tables were on par with file directory information (a little like VTOC ttr). I had also made the change to "page" less-critical portions of the fixed kernel (both reducing fixed storage requirements). Just had to have the dependency graph ... so there wasn't deadly embrace ... where something had been paged out that was required in order to perform restore/page-in.
note that 3090 put something like pagefile in RAM .... called expanded store. basically the processor memory technology packaging couldn't get all of the storage they wanted within latency for processor cache miss. there had been "electronic" (paging) disks (prior to 3090) where additional memory was used to simulate fixed-head disks.
3090 expanded store was an extra wide bus with a synchronous move instruction (push pages out to expanded store or pull pages back in from expanded store) ... the elapsed time trade-off for the synchronous operation was significantly shorter than the pathlength elapsed time for performing asynchronous i/o operation.
for later machines, memory technology packaging enabled as much memory as you wanted for processor use ... but "LPAR" configuration (still) allowed some of it to be allocated for "expanded store" simulation. all other things being equal ... directly being able to always access a page is more efficient than moving it back&forth. However, there were some system idiosyncrasies (and/or bugs) ... that allowed configurations to perform better with some "expanded store" ... rather than everything directly addressable.
A simple scenario was system limitations of 2gbyte addressability and real memory much larger than 2gbyte.
The 3090 expanded store bus was also used for attaching HIPPI channels/devices (100mbyte/sec) ... since standard 3090 i/o facilities couldn't operate that fast. HIPPI was cut into the side of the expanded store bus and channel I/O was performed with peek/poke paradigm (using "reserved" expanded store addresses).
misc. past posts mentioning expanded store:
https://www.garlic.com/~lynn/2003p.html#41 comp.arch classic: the 10-bit byte
https://www.garlic.com/~lynn/2006.html#17 {SPAM?} DCSS as SWAP disk for z/Linux
https://www.garlic.com/~lynn/2006k.html#57 virtual memory
https://www.garlic.com/~lynn/2006l.html#43 One or two CPUs - the pros & cons
https://www.garlic.com/~lynn/2006r.html#36 REAL memory column in SDSF
https://www.garlic.com/~lynn/2006s.html#16 memory, 360 lcs, 3090 expanded store, etc
https://www.garlic.com/~lynn/2006s.html#17 bandwidth of a swallow (was: Real core)
https://www.garlic.com/~lynn/2006s.html#20 real core
https://www.garlic.com/~lynn/2007c.html#23 How many 36-bit Unix ports in the old days?
https://www.garlic.com/~lynn/2007f.html#18 What to do with extra storage on new z9
https://www.garlic.com/~lynn/2007g.html#67 Unusual Floating-Point Format Remembered?
https://www.garlic.com/~lynn/2007o.html#26 Tom's Hdw review of SSDs
https://www.garlic.com/~lynn/2007p.html#11 what does xp do when system is copying
https://www.garlic.com/~lynn/2007s.html#9 Poster of computer hardware events?
https://www.garlic.com/~lynn/2008.html#49 IBM LCS
https://www.garlic.com/~lynn/2008b.html#15 Flash memory arrays
https://www.garlic.com/~lynn/2008f.html#6 Fantasy-Land_Hierarchal_NUMA_Memory-Model_on_Vertical
https://www.garlic.com/~lynn/2008f.html#8 Fantasy-Land_Hierarchal_NUMA_Memory-Model_on_Vertical
https://www.garlic.com/~lynn/2010.html#86 locate mode, was Happy DEC-10 Day
https://www.garlic.com/~lynn/2011e.html#62 3090 ... announce 12Feb85
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: Z chip at ISSCC Newsgroups: comp.arch Date: Thu, 28 Apr 2011 16:31:55 -0400Robert Myers <rbmyersusa@gmail.com> writes:
It was one of the reasons that we did cluster scale-up with large number
of processors ... misc. old email mentioning cluster scale-up
https://www.garlic.com/~lynn/lhwemail.html#medusa
later there was somerset/aim ... with apple & motorola (the executive we
reported to when doing cluster scale-up ... went over to head up
somerset) ... which did some work on cache consistency. misc. old
email 801, risc, romp, rios, somerset, etc
https://www.garlic.com/~lynn/lhwemail.html#801
misc. past posts mentioning future system
https://www.garlic.com/~lynn/submain.html#futuresys
misc. past posts mentioning cluster stuff
https://www.garlic.com/~lynn/subtopic.html#hacmp
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: how to get a command result without writing it to a file Newsgroups: comp.lang.rexx Date: Thu, 28 Apr 2011 18:12:24 -0400"Dave Saville" <dave@invalid.invalid> writes:
the science center had done virtual machine operating system called cp40 on a specially modified 360/40 with virtual memory hardware. when standard 360/67 (w/virtual memory) became available, the 360/40 was replaced with 360/67 and cp40 morphed into cp67.
3 people from the science center came out to univ (where i was
undergraduate) in jan68 to install cp67 (the univ. had gotten a 360/67
for running tss/360 .... which wasn't going well so it was running
360/67 in real address mode with os/360). misc. past posts mentioning
science center
https://www.garlic.com/~lynn/subtopic.html#545tech
while at univ I rewrote significant sections of cp67 ... lots of
enhancements to pathlength, resource management, page replacement, i/o
operations, etc. old post with part of presentation I have at '68 share
meeting later that year (mostly about pathlength changes):
https://www.garlic.com/~lynn/94.html#18 CP/67 & OS MFT14
i re-arranged the kernel so that all the routines to be paged were located contiguously at the end of the kernel and organized into 4kbyte page chunks. I created a dummy virtual memory table that represented the kernel image and drew the line at the start of "pageable" kernel ... and allowed all of those pages to be "paged". I then modified the routine that handled kernel calls/returns ... so that addresses above the "line" would be page fetched & locked (using convention for doing simulated virtual machine I/O) ... and returns would decrement the lock count.
for virtual machine ... there were two types of tables for each virtual page. I created another dummy virtual memory (similar to what I did for paging kernel) for each virtual machine ... where I would page virtual machine control information. Each virtual machine virtual segment had a pagetable and a "swaptable" ... the "swaptable" contained control information about each virtual page ... the location of the virtual page on disk, shadow image of the virtual page storage keys, and some misc. other stuff. when all virtual pages for a segment were no longer in storage, i could disolve the pagetable (using the segment invalid flag), and page-out the corresponding swaptable.
some misc. old email about page replacement algorithm
https://www.garlic.com/~lynn/lhwemail.html#globallru
some old email about porting changes from cp67 to vm370
https://www.garlic.com/~lynn/2006v.html#email731212
https://www.garlic.com/~lynn/2006w.html#email750102
https://www.garlic.com/~lynn/2006w.html#email750430
part of the support mentioned above ... was moving cms paged-mapped
filesystem implementation from cp67 to vm370. misc. past posts
mentioning cms paged-mapped filesystem support
https://www.garlic.com/~lynn/submain.html#mmap
at various points, some of the stuff would periodically leak out into various product releases.
some recent posts about paging, fixed-head devices, electronic disks
https://www.garlic.com/~lynn/2011f.html#56 Drum Memory with small Core Memory?
https://www.garlic.com/~lynn/2011f.html#59 Drum Memory with small Core Memory?
https://www.garlic.com/~lynn/2011f.html#61 Drum Memory with small Core Memory?
slightly bring it back to rexx ... very early days of rexx (when it was
still called rex) ... i wanted to demonstrate that rex wasn't just
another pretty scripting language. For the demo, I would rewrite the
problem-determination/dump-analyzer application implemented in assembler
(had a whole dept. in endicott supporting it) in rex ... doing it in
half-time over period of 3months and it would have ten times the
function and run ten times faster. misc. past posts mentioning dumprx
https://www.garlic.com/~lynn/submain.html#dumprx
for other random ... from old dmsrex h-assembler listing
DMSREX assembled from REXPAK (2099 records, 04/15/83 17:19:55)
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: program coding pads Newsgroups: alt.folklore.computers Date: Fri, 29 Apr 2011 18:27:10 -0400Charles Richmond <frizzle@tx.rr.com> writes:
misc. past posts getting to play disk engineer
https://www.garlic.com/~lynn/subtopic.html#disk
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 30 Apr, 2011 Subject: Wylbur, Orvyl, Milton, CRBE/CRJE were all used (and sometimes liked) in the past. Blog: Old Geek Registryre:
do you mean this one:
http://www.stanford.edu/dept/its/support/wylorv/
this has little ORVYL history ... apparently having been done on
360/67 (360/65 with virtual memory hardware) at stanford in the late
60s (similar to what Univ. of Mich. did for MTS on their 360/67)
... see section 1.3:
http://www.slac.stanford.edu/spires/explain/manuals/ORVMAN.HTML
above mentions Richard Carr did substantial work on ORVYL-II (possibly migration from 360/67 to 370 w/virtual memory support).
In the early 80s, Carr was at Tandem and also finishing up his PHD at stanford in global LRU page replacement algorithm (and encountering significant academic resistance from some corners regarding global LRU vis-a-vis local LRU).
This is old mail
https://www.garlic.com/~lynn/2006w.html#email821019
trying to provide supporting assistance in this post
https://www.garlic.com/~lynn/2006w.html#46
Jim Gray had approached me at SIGOPS (Asilomar, dec81) about providing supporting evidence ... since I had done something similar as undergraduate in the 60s on cp67. Even tho the references was to work I had done as undergraduate ... it took me nearly a year to get corporate approval to send the reply (hopefully it wasn't because corporation taking sides in the academic dispute).
additional global LRU topic drift ... old email
https://www.garlic.com/~lynn/lhwemail.html#globallru
recent MTS post in a.f.c. with some URLs to old (MTS) pictures/lore
https://www.garlic.com/~lynn/2011f.html#63
another recent post in a.f.c. mentioning having hacked 2741&tty
terminal support into HASP with a conversational editor that
implemented CMS editor syntax (rewrite from scratch) ... also
references other HASP, Future System, SVS & MVS history
https://www.garlic.com/~lynn/2011d.html#73
wiki reference (for orvyl, wylbur and milton):
https://en.wikipedia.org/wiki/ORVYL_and_WYLBUR
refers to SLAC & CERN continued to use into the 90s. SLAC & CERN were sister organizations sharing lots of software. SLAC also was host of monthly (vm/cms user group) BAYBUNCH meetings.
for other topic drift ... reference to (vm/cms) SLAC being 1st
webserver outside CERN
https://ahro.slac.stanford.edu/wwwslac-exhibit
story of SGML involving into HTML at CERN
http://infomesh.net/html/history/early/
GML was invented in 1969 ... misc. past posts
https://www.garlic.com/~lynn/submain.html#sgml
at the science center ... misc. past posts
https://www.garlic.com/~lynn/subtopic.html#545tech
before morphing into ISO standard SGML in the late 70s (and more than a decade later morphing into HTML).
Original Orvyl was operating system for "bare machine" (360/67 had virtual memory hardware suppot) ... later Orvyl appears to have been redone for hosting under MVS (more like CICS handling all its own subtasks?). Various flavors of Wylbur were also done that could run w/o Orvyl.
This is recent post about Adventure ... which appeared to have come
over from Stanford's PDP machine to Stanford's Orvyl (& redone from
PDP fortran to PLI) which could run in standard TSO environment (as
well as CMS OS/360 simulation) with the asm code (converted
tread/twrite to tget/tput):
https://www.garlic.com/~lynn/2011b.html#41
In 70s, I had gotten CMS Fortran version ... which appeared to have come over from Stanford's PDP machine to TYMSHARE's PDP machine and then to TYMSHARE vm370/cms.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: The IBM Selective Sequence Electronic Calculator Newsgroups: alt.folklore.computers Date: Sun, 01 May 2011 09:35:57 -0400jmfbahciv <See.above@aol.com> writes:
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 2 May, 2011 Subject: Wylbur, Orvyl, Milton, CRBE/CRJE were all used (and sometimes liked) in the past. Blog: Old Geek Registryre:
HONE (HONE provided online worldwide sales&marketing support, the US datacenter's had been consolidated in Palo Alto in the mid-70s in bldg. that is now next door to Facebook bldg), SLAC, Tymshare, Stanford were all in close physical proximity. I would go by some number ... especially on the day of the monthly Baybunch meetings at SLAC.
Tymshare had started hosting the online VMSHARE computer conferencing
Au76 ... archives here:
http://vm.marist.edu/~vmshare/
This is old email about dropping by Tymshare and getting a demo of
Adventure ... and then making arrangements for source. However,
somebody in the UK walked a copy from a univ. machine over to a
corporate machine and sent me a copy over the internal network
https://www.garlic.com/~lynn/2006y.html#email780405
https://www.garlic.com/~lynn/2006y.html#email780405b
The internal network was larger than the arpanet/internet from just
about the beginning until late '85 or possibly early '86
... misc. past posts
https://www.garlic.com/~lynn/subnetwork.html#internalnet
I had also made provisions with Tymshare to get monthly copies of the
VMSHARE (and later PCSHARE) files and made them available on the
internal network ... misc. old email mentioning VMSHARE
https://www.garlic.com/~lynn/lhwemail.html#vmshare
I also got blamed for online computer conferencing on the internal network in the late 70s and early 80s.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: PIC code, RISC versus CISC Newsgroups: comp.arch Date: Mon, 02 May 2011 11:04:58 -0400Michael S <already5chosen@yahoo.com> writes:
I had done page-mapped filesystem for cp67/cms (and ported to vm370/cms)
some past posts
https://www.garlic.com/~lynn/submain.html#mmap
with semantics that allowed for "PIC" (arbitrary location for both code
& data) ... however, much of cms adopted os/360 applications and
conventions (using some amount of simulation of os/360 functions). as a
result, it prevented effective use of "PIC" w/o a lot of hacking
... misc. past posts mentioning painful hacking to achieve "PIC" with
code that was predominately os/360 oriented
https://www.garlic.com/~lynn/submain.html#adcon
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: Overloaded acronyms Newsgroups: bit.listserv.ibm-main Date: 2 May 2011 08:40:19 -0700at one point there was almost PCO (personal computing option) ... sort of TSO for VS/1 ... however it was eventually pointed out that PCO was also initials for political party in europe ... and PCO morphed into VS/PC.
there was one plan to have VS/1 machines already preloaded with vm/cms (sort of like early flavor of LPARs) ... and using CMS as the interactive component. PCO was being positioned as an alternative. The PCO group had a "simulator" showing PCO performance something like ten times that of vm/cms ... their simulation group would "run" some number of "benchmarks" ... and then the vm/cms group were asked to perform similar (real) benchmarks (taking up a significant percentage of all vm/cms resources on the benchmarks ... taken away from doing actual development). When PCO was finally operational ... it turned out to be slower than vm/cms (but they managed to waste a significant percentage of vm/cms development resources on the fictitious benchmarks)
various internal politics blocks the strategy to preload vm/cms on every
mid-range machine. then in the wake of the death of Future System effort
... the MVS/XA effort managed to convince corporate to completely
kill-off vm/cms (shutting down the development group and moving
everybody to POK to support MVS/XA ... with claim that they wouldn't
otherwise make MVS/XA ship schedule). Endicott eventually managed to
save the vm/cms product mission ... but had to reconstitute a
development group from scratch. misc. past posts mentioning future system
https://www.garlic.com/~lynn/submain.html#futuresys
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 2 May, 2011 Subject: Wylbur, Orvyl, Milton, CRBE/CRJE were all used (and sometimes liked) in the past. Blog: Old Geek Registryre:
??? ibm-main mailing list started on bitnet in the 80s ... post of
mine from a couple yrs ago
http://www.mail-archive.com/ibm-main@bama.ua.edu/msg79154.html
also archived here:
https://www.garlic.com/~lynn/2008k.html#85
bitnet wiki page
https://en.wikipedia.org/wiki/BITNET
using listserv software
http://www.lsoft.com/corporate/history_listserv.asp
started on EARN (version of bitnet in europe). old email about getting
EARN going
https://www.garlic.com/~lynn/2001h.html#email840320
past email mentioning bitnet &/or earn
https://www.garlic.com/~lynn/subnetwork.html#bitnet
listserv was subset of function of the (earlier) toolrun facility
developed internally for the internal network ... misc. past posts
mentioning internal network
https://www.garlic.com/~lynn/subnetwork.html#internalnet
I had been blamed for online computer conferencing on the internal network in the late 70s and early 80s. Folklore is that when the executive committee (chairman, ceo, pres, etc) was informed of online computer conferencing (and the internal network), 5out6 wanted to fire me.
From IBM Jargon
Tandem Memos - n. Something constructive but hard to control; a fresh
of breath air (sic). That's another Tandem Memos. A phrase to worry
middle management. It refers to the computer-based conference (widely
distributed in 1981) in which many technical personnel expressed
dissatisfaction with the tools available to them at that time, and
also constructively criticized the way products were [are]
developed. The memos are required reading for anyone with a serious
interest in quality products.
... snip ...
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 3 May, 2011 Subject: DCSS ... when shared segments were implemented in VM Blog: z/VMre:
recent thread in comp.arch with topic drift on "PIC code" (aka
"position independent") ... the base implementation (from which DCSS
took a small subset) allowed for position independent:
http://groups.google.com/group/comp.arch/browse_thread/thread/45305e22397b01de/8a4fc64796459cd?q=group:comp.arch+insubject:PIC#08a4fc64796459cd
my post in thread:
tss/360 supported "PIC" for 360/67 ... but os/360 conventions didn't.
I had done page-mapped filesystem for cp67/cms (and ported to
vm370/cms) some past posts
https://www.garlic.com/~lynn/submain.html#mmap
with semantics that allowed for "PIC" (arbitrary location for both
code & data) ... however, much of cms adopted os/360 applications and
conventions (using some amount of simulation of os/360 functions). as
a result, it prevented effective use of "PIC" w/o a lot of hacking
... misc. past posts mentioning painful hacking to achieve "PIC" with
code that was predominately os/360 oriented
https://www.garlic.com/~lynn/submain.html#adcon
....
also archived here
https://www.garlic.com/~lynn/2011f.html#76
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: TSO Profile NUM and PACK Newsgroups: bit.listserv.ibm-main Date: 3 May 2011 10:14:48 -0700PaulGBoulder@AIM.COM (Paul Gilmartin) writes:
at the univ., I was making so many cp67/cms source changes that I created a pre-processor for update ... that added extra field to the insert&replace control statements (aka "$") that would generate sequence numbers of the new lines (otherwise they had to be manually entered/typed).
later in early 70s ... there was large "exec" wrapper that supported multiple updates in specified sequence. there was a joint development effort with endicott that added 370 virtual machine simulation to cp67 (that ran on 360/67) ... including new instructions and virtual memory support that had several differences from 360/67.
There was "base" set of local enhancements to cp67 ... referred to as the "L" updates ... then could apply the "H" updates to provide option for 370 virtual machines (in addition to 360 virtual machines), and then could apply the "I" updates which modified cp67 to run on 370 machine (rather than 360/67).
"cp67i" was running regularly in 370 virtual machine for a year before the first 370 engineering machine with virtual memory hardware support was operational (a 370/145 in endicott) ... in fact, booting "cp67i" on the engineering machine was part of early validation test for the hardware. turns out boot failed ... because of "errors" in the hardware implementation (cp67i was quickly patched to correspond with incorrect hardware ... and then booted successfully).
By the time of vm370/cms, the multi-level update conventions ... had been incorporated into CMS update command (drastically reducing the exec wapper) and various editors. Editors were also modified to have option to generate edit session saved changes as an incremental update file (as opposed to replacing the original file with the changes).
There is folklore about HASP/JES2 group had moved to cms source development process ... which resulted in various kinds of problems for exporting into standard POK product release environment.
In the mid-80s, Melinda had requested anybody with a copy of the original cp67/cms multi-level update implementation. It turns out that I had complete set on archived tapes in the Almaden datacenter tape library. Her request was timely since a couple months later the Almaden datacenter had an operations problem with mounting random tapes as scratch (destroying large number of tapes, including ones with my archived info from the 70s ... in some cases multiple tapes with replicated copies ... including those with large amount of cp67/cms files).
old email exchange with Melinda
https://www.garlic.com/~lynn/2006w.html#email850906
https://www.garlic.com/~lynn/2006w.html#email850906b
https://www.garlic.com/~lynn/2006w.html#email850908
Melinda's home page has moved:
https://www.leeandmelindavarian.com/Melinda#VMHist
I had done kindle conversion of her history ... which she now has up:
https://www.leeandmelindavarian.com/Melinda/neuvm.azw
cms update command reference:
http://publib.boulder.ibm.com/infocenter/zos/v1r10/topic/com.ibm.zos.r10.asmk200/ap5cms8.htm
xedit cms command reference (including mention of update option support)
http://publib.boulder.ibm.com/infocenter/zvm/v5r4/topic/com.ibm.zvm.v54.dmsb6/xco.htm
note that univ source update uses "down-dates" ... i.e. the "current" source file includes all changes ... but there are history files that allows changes to be "regressed" to earlier versions. the cms "up-dates" process would freeze the original source (for some period of time) and have sequence of incremental source updates that would be applied in sequence to arrive at most up-to-date file to be compiled/assembled.
misc. past posts mentioning cms update command:
https://www.garlic.com/~lynn/2002g.html#67 Coulda, Woulda, Shoudda moments?
https://www.garlic.com/~lynn/2002n.html#39 CMS update
https://www.garlic.com/~lynn/2002p.html#2 IBM OS source code
https://www.garlic.com/~lynn/2003e.html#38 editors/termcap
https://www.garlic.com/~lynn/2003e.html#66 History of project maintenance tools -- what and when?
https://www.garlic.com/~lynn/2003f.html#1 History of project maintenance tools -- what and when?
https://www.garlic.com/~lynn/2003k.html#47 Slashdot: O'Reilly On The Importance Of The Mainframe Heritage
https://www.garlic.com/~lynn/2004b.html#59 A POX on you, Dennis Ritchie!!!
https://www.garlic.com/~lynn/2004g.html#43 Sequence Numbbers in Location 73-80
https://www.garlic.com/~lynn/2005f.html#44 DNS Name Caching
https://www.garlic.com/~lynn/2005i.html#30 Status of Software Reuse?
https://www.garlic.com/~lynn/2005p.html#45 HASP/ASP JES/JES2/JES3
https://www.garlic.com/~lynn/2006n.html#45 sorting
https://www.garlic.com/~lynn/2006o.html#19 Source maintenance was Re: SEQUENCE NUMBERS
https://www.garlic.com/~lynn/2006o.html#21 Source maintenance was Re: SEQUENCE NUMBERS
https://www.garlic.com/~lynn/2006u.html#26 Assembler question
https://www.garlic.com/~lynn/2006w.html#42 vmshare
https://www.garlic.com/~lynn/2007b.html#7 information utility
https://www.garlic.com/~lynn/2009h.html#48 Book on Poughkeepsie
https://www.garlic.com/~lynn/2009s.html#17 old email
https://www.garlic.com/~lynn/2009s.html#37 DEC-10 SOS Editor Intra-Line Editing
https://www.garlic.com/~lynn/2010k.html#13 Idiotic programming style edicts
https://www.garlic.com/~lynn/2011c.html#3 If IBM Hadn't Bet the Company
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: TSO Profile NUM and PACK Newsgroups: bit.listserv.ibm-main Date: 3 May 2011 11:45:35 -0700re:
Note: UNIX traces some of its history back to CTSS by way of MULTICS
done on 5th flr of 545 tech sq. VM370/CMS also traces history back to
CTSS by way of CP67/CMS and CP40/CMS done (at science center) on 4th flr
of 545 tech sq. misc. past posts mentioning science center
https://www.garlic.com/~lynn/subtopic.html#545tech
example is UNIX (document formating) runoff/roff looks very much like
CTSS "runoff". original CMS (document formating) "script" also looked
very much like CTSS "runoff" ... this was before GML was invented at the
science center in 1969 and GML tag processing added to CMS
"script". misc. past posts mentioning GML &/or SGML
https://www.garlic.com/~lynn/submain.html#sgml
CTSS reference:
https://www.multicians.org/thvv/7094.html
Discusses some of CTSS relationship to CP/CMS, MULTICS, and UNIX
(mentions that TSO is in no way related):
https://en.wikipedia.org/wiki/Compatible_Time-Sharing_System
Discusses some of CP/CMS relationship to CTSS
https://en.wikipedia.org/wiki/History_of_CP/CMS
Multics reference
https://www.multicians.org/general.html
Unix and Multics reference
https://www.multicians.org/unix.html
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 03 May, 2011 Subject: Bank email archives thrown open in financial crash report Blog: Financial Crime Risk, Fraud and Securityre:
The visible hand; In MIT talk, Eliot Spitzer defends role of
government in regulating markets, claims economy still 'on the
precipice' of deep problems.
http://web.mit.edu/newsoffice/2011/spitzer-talk-0428.html
Did Deutsche Bank Really Hide Its Bad Audits In A Closet?
http://blogs.forbes.com/halahtouryalai/2011/05/03/did-deutsche-bank-really-hide-its-bad-audits-in-a-closet/
Justice Department, SEC Probing Senate Findings on Goldman
http://www.bloomberg.com/news/2011-05-03/levin-report-accusing-goldman-of-deception-referred-to-u-s-justice-sec.html
TV business news hitting the Goldman item this morning ... somewhat about whether SEC will actually follow through with anything significant about Goldman (in contrast to doing very little for the past decade).
There are also some references that house may be able to significantly
hinder SEC from doing anything by cutting its budget (regardless of
what senate is trying to accomplish)
http://dealbook.nytimes.com/2011/05/03/u-s-regulators-face-budget-pinch-as-mandates-widen/
In the Madoff hearings, the person that had tried unsuccessfully for a decade to get SEC to do anything about Madoff mentioned that part of SEC problem was that they are all lawyers with little training in financial forensics (no skill in creating a case). He was also asked about the need for new regulation. The comment was that while new regulation may be needed, much more important was transparency and visibility (that regulation and enforcement is extremely more difficult ... and costs go through the roof when there isn't transparency and visibility). This is similar to other testimony about costs/regulation going skyhigh when business processes are misaligned and people are motivated to do the wrong thing. Changing the environment and providing transparency and visibility drastically reduce the cost and amount of regulation.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: program coding pads Newsgroups: alt.folklore.computers Date: Wed, 04 May 2011 10:20:18 -0400ArarghMail105NOSPAM writes:
reference to after a move (not long after above post) ... and not being
able to get either cable or dsl ... being told that they would be happy
to install T1 frame-relay at $1200/month.
https://www.garlic.com/~lynn/2008d.html#94 The Economic Impact of Stimulating Broadband Nationally
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: program coding pads Newsgroups: alt.folklore.computers Date: Wed, 04 May 2011 11:08:56 -0400scott@slp53.sl.home (Scott Lurndal) writes:
I frequently see mbyte/sec download ... although 600kbyte/sec is more normal (but it is hard to see whether it is constrained at my end ... or the server is hitting load) ... all for $33/month ... compared to $1200/month for 150kbyte/sec download. T1 works out to little less than $1000/month/100kbyte ... compared to cable at $5/month/100kbyte.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@GARLIC.COM (Anne & Lynn Wheeler) Subject: Re: SV: USS vs USS Newsgroups: bit.listserv.ibm-main Date: 4 May 2011 15:01:28 -0700mike.a.schwab@GMAIL.COM (Mike Schwab) writes:
folklore is that after demise of future system, some of the participants retreated to rochester and did s/38 ... which then morphs into as/400
in the 80s, tss/370 got something of a new life ... as base for special bid mainframe unix for AT&T ... stripped down tss/370 kernel (SSUP) with AT&T doing unix interfaces to the SSUP kernel interface (in some sense this is somewhat analogous to USS for MVS). this was competing with Amdahl's GOLD/UTS unix internally inside AT&T.
AIX/370 (in conjunction with AIX/386) was done by palo alto group using the unix-like LOCUS done at UCLA. This was similar but different from the unix-like MACH done at CMU ... which was used by a number of vendors including NeXT and morphs into current Apple operating system after Jobs returns to Apple. AIX/370 morphs into AIX/ESA.
The "argument" for (Amdahl) UTS under vm370, aix/370 under vm370, tss/370 ssup, and vm/ix (on vm370) was that the cost to add mainframe RAS&erep to unix was several times larger than the base, direct, straight-forward unix port (running under vm370 &/or tss/370 leveraged the already existing ras&erep support w/o having to re-implement directly in unix). This was aggravated by field service stand that it wouldn't service/support machines that lacked mainframe RAS&erep.
I ran internal advanced technology conference in '82 ... and some of the
presentation were about VM/IX implementation ... old post reference:
https://www.garlic.com/~lynn/96.html#4a
Palo Alto group had also been working with Berkeley to port their unix-like BSD to mainframe ... but they got redirected instead doing a PC/RT port ... released from ACIS as "AOS" ... as an alternative UNIX to the "official" AIXV2.
The wiki page says much of the AIX v2 kernel was written in PL/I. The
issue was that the original "displaywriter" was based on ROMP, cp.r, and
PL.8 (sort of pli subset). Redirected to the unix workstation market
required unix&C (all being done by the company that had done pc/ix and
had been involved in vm/ix). For the internal people, a project called
VRM was devised ... a sort of abstract virtual machine layer ... to be
done by the internal employees trained in pl.8. The claim was that the
combination VRM plus unix port to VRM ... could be done in shorter time
and less resources than unix port directly to ROMP hardware. The exact
opposite was shown when the palo alto group did the BSD port direct to
ROMP hardware (for "AOS"). VRM+unix drastically increased original/total
development costs, life-cycle support costs and complicated things like
new device drivers (since both non-standard unix/c device driver to VRM
interface as well as VRM/pl.8 device driver had to be developed &
supported). misc. past posts mentioning 801, romp, rios, pc/rt, aixv2,
aixv3, power, rs/6000, etc
https://www.garlic.com/~lynn/subtopic.html#801
misc. old email mentioning 801
https://www.garlic.com/~lynn/lhwemail.html#801
Besides various other issues, the AIX wiki page skips over the whole
generation of OSF
https://en.wikipedia.org/wiki/Open_Software_Foundation
and the "unix wars"
https://en.wikipedia.org/wiki/UNIX_wars
Project Monterey
https://en.wikipedia.org/wiki/Project_Monterey
skips over the whole cluster scale-up after IBM bought Sequent and
support for Sequent's 256-way SCI-based Numa-Q. Recent posts in
(linkedin) "Greater IBM" (current & former IBMer) discussion
https://www.garlic.com/~lynn/2011d.html#7 IBM Watson's Ancestors: A Look at Supercomputers of the Past
the sequent wiki ... mentioned in the above post ... used to be somewhat
more caustic about sequent being dropped shortly after the sponsoring
executive retired:
https://en.wikipedia.org/wiki/Sequent_Computer_Systems
as noted in the "Greater IBM" post ... at one time, IBM had been providing quite a bit of funding for Chen's Supercomputer ... Sequent later acquires Chen Supercomputer and Chen becomes CTO at Sequent ... we do some consulting for Chen (before Sequent purchase by IBM).
Part of the speculation for IBM's purchase of Sequent was that Sequent was major platform for some of the IBM mainframe simulator products.
much of the "posix" (aka unix) support in MVS during the first half of the 90s was sponsored by the head of the disk division software group. in the late 80s, a senior disk engineer got a talk scheduled at the internal, annual, world-wide communication group conference ... and opened the talk with the statement that the communication group was going to be responsible for the demise of the disk division (because the stranglehold that the communication group had on datacenters). Large amounts of data was fleeing datacenters to more distributed computing friendly platforms. The disk division had attempted to come out with traditional products to address the problem ... but they were constantly blocked by the communication group. As a result, there were doing all sorts of things "outside-the-box" to try and work around the communication group's roadblocks. the head of the disk division software group would periodically ask us to consult on some of the efforts.
for other drift, recent thread in comp.arch about tss/360 supported
"position independent code" (i.e. possible to directly map a disk
image into virtual memory at any arbitrary virtual address w/o having to
perform any link-edit modifications to the contents of that image)
... and the horrendous problems attempting to do anything similar using
anything from the os/360 genre:
https://www.garlic.com/~lynn/2011f.html#76 PIC code, RISC versus CISC
also referenced:
https://www.garlic.com/~lynn/2011f.html#79 DCSS ... when shared segments were implemented in VM
misc. past posts mentioning (tss/370) ssup:
https://www.garlic.com/~lynn/2004q.html#37 A Glimpse into PC Development Philosophy
https://www.garlic.com/~lynn/2005b.html#13 Relocating application architecture and compiler support
https://www.garlic.com/~lynn/2005c.html#20 [Lit.] Buffer overruns
https://www.garlic.com/~lynn/2005d.html#61 Virtual Machine Hardware
https://www.garlic.com/~lynn/2005s.html#34 Power5 and Cell, new issue of IBM Journal of R&D
https://www.garlic.com/~lynn/2006f.html#26 Old PCs--environmental hazard
https://www.garlic.com/~lynn/2006g.html#2 The Pankian Metaphor
https://www.garlic.com/~lynn/2006m.html#30 Old Hashing Routine
https://www.garlic.com/~lynn/2006p.html#22 Admired designs / designs to study
https://www.garlic.com/~lynn/2006t.html#17 old Gold/UTS reference
https://www.garlic.com/~lynn/2007.html#38 How many 36-bit Unix ports in the old days?
https://www.garlic.com/~lynn/2007b.html#3 How many 36-bit Unix ports in the old days?
https://www.garlic.com/~lynn/2007k.html#43 John W. Backus, 82, Fortran developer, dies
https://www.garlic.com/~lynn/2007m.html#69 Operating systems are old and busted
https://www.garlic.com/~lynn/2008e.html#1 Migration from Mainframe to othre platforms - the othe bell?
https://www.garlic.com/~lynn/2008e.html#49 Any benefit to programming a RISC processor by hand?
https://www.garlic.com/~lynn/2008l.html#82 Yet another squirrel question - Results (very very long post)
https://www.garlic.com/~lynn/2008r.html#21 What if the computers went back to the '70s too?
https://www.garlic.com/~lynn/2010c.html#43 PC history, was search engine history, was Happy DEC-10 Day
https://www.garlic.com/~lynn/2010e.html#17 Senior Java Developer vs. MVS Systems Programmer (warning: Conley rant)
https://www.garlic.com/~lynn/2010e.html#72 Entry point for a Mainframe?
https://www.garlic.com/~lynn/2010i.html#28 someone smarter than Dave Cutler
https://www.garlic.com/~lynn/2010o.html#0 Hashing for DISTINCT or GROUP BY in SQL
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 04 May, 2011 Subject: Bank email archives thrown open in financial crash report Blog: Financial Crime Risk, Fraud and Securityre:
also in the area of transparency and visibility: The Dodd-Frank Act
will demand data transparency:
http://searchdatamanagement.techtarget.com/news/2240035386/The-Dodd-Frank-Act-could-mean-a-data-management-mess-for-some
one of the comments in the above ... is that for some institutions the change for transparency is going to make Y2K look like a walk-in-the-park ... some rhetoric that is effectively an onerous penalty for these institutions (akin to comments about how onerous the sarbanes-oxley audits were suppose to be ... but it turns out it was more like a reward for the audit industry for having done enron ... since SEC didn't do anything ... apparently prompting GAO to do reports finding uptick in public company fraudulent financial fillings). with regard to Y2K ... there is folklore that one of the too-big-to-fail institutions had outsourced to the lowest bidder ... later finding it was a front for a ethnic organized crime operation ... and the code containing all sorts of clandestine wire transfers
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 6 May, 2011 Subject: Gee... I wonder if I qualify for "old geek"? Blog: Old Geek Registryre:
They let me wander around the san jose plant site ... even though I was in bldg. 28. One of the things I noticed was that the development dasd "testcells" were running scheduled "stand-alone" testing on various mainframes. They had tried putting up MVS on the mainframes to allow multiple, concurrent testing, on-demand testing ... but found MVS had 15min MTBF in that environment (system hung and/or failed requiring reboot).
I offered to rewrite I/O supervisor to make it bullet proof and never fail ... so they could have on-demand, concurrent testing at anytime, significantly improving productivity. A side effect was that the mainframes in bldg. 14&15 became available for other uses also ... since the testcell testing consumed very little of processor resources.
Now, bldg. 28 had a 370/195 running under MVT ... but even high priority work could have several week turn-around. One of the applications was "air-bearing simulation" for design of disk "floating heads". Bldg. 15 got one of the first engineering 3033s for disk testing. The 370/195 had peak thruput around 10mips ... but normal codes ran closer to 5mips. By comparison the 3033 ran around 4.5mips (basically 1.5 times 168-3). All the disk testing consumed a percent or two of the 3033 ... so there was lots of idle cycles to do other things. One of the things setup was to get the air-bearing simulation that ran on 370/195 MVT system (with several week turn-around) executing on the 3033 in bldg. 15 (where they could possibly get several turn-arounds a day).
Later, after STL was getting production 3033s ... there was
investigation about also being able to run the air-bearing simulation
on their offshift (effectively idle) 3033s ... old email from long ago
and far away (although stl datacenter had traditional "billing"
program which had to be handled):
Date: 05/29/80 17:28:21
To: wheeler
xxxxxx says he sees no reason why the air bearing problem couldn't run
on third shift STLVM1 or 3 ... both 3033s and both now up on third
shift according to him. If true, guess that leaves the settling on the
bill....
... snip ... top of post, old email index
The bldg. 14&15 mainframe use was totally off any datacenter
"accounting" grid ... since they were theoretically "stand-alone"
testing machines. misc. past posts mentioning getting to play disk
engineer in bldgs. 14&15
https://www.garlic.com/~lynn/subtopic.html#disk
Date: 06/04/80 18:06:51
From: wheeler
I've been asked to look at the Air Bearing program tomorrow to see if
I can identify anthing for performance improvements.
... snip ... top of post, old email index
Date: 06/05/80 07:53:36
From: wheeler
re: air bearing; Checking pli manual, ENV(TOTAL) can't be used with
VBS records, almost anything else but VBS. I forget, but what is O/S
close option/equivalent to CMS TCLOSE, i.e. checkpoint file, but leave
it open'ed. The only other two that come to mind 'quickly' is REORDER
in PLI, and specifying FORTRAN and redo some of the program as FORTQ
subroutine.
... snip ... top of post, old email index
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 06 May, 2011 Subject: Court OKs Firing of Boeing Computer-Security Whistleblowers Blog: Financial Crime Risk, Fraud and SecurityCourt OKs Firing of Boeing Computer-Security Whistleblowers
from above:
Two Boeing auditors were legally fired after they exposed to the press
internal documents suggesting the aerospace and military contractor
lacked computer-security safeguards, a federal appeals court ruled
Tuesday.
... snip ...
also
"Members of the media are not included."
https://financialcryptography.com/mt/archives/001313.html
In the past I had commented that possibly the only significant part of SOX was about informants. In the congressional Madoff hearings, the person that had unsuccessfully tried for a decade to get SEC to do something about Madoff, commented that tips turn up 13 times more fraud than audits (and SEC didn't have a tip line ... but did have 1-800 line to complain about audits)
misc. past posts mentioning tips & audits:
https://www.garlic.com/~lynn/2011.html#46 What do you think about fraud prevention in the governments?
https://www.garlic.com/~lynn/2011.html#48 What do you think about fraud prevention in the governments?
https://www.garlic.com/~lynn/2011e.html#56 In your opinon, what is the highest risk of financial fraud for a corporation ?
https://www.garlic.com/~lynn/2011f.html#62 Mixing Auth and Non-Auth Modules
--
virtualization experience starting Jan1968, online at home since Mar1970
From: Anne & Lynn Wheeler <lynn@garlic.com> Subject: Re: The first personal computer (PC) Newsgroups: alt.folklore.computers Date: Fri, 06 May 2011 13:06:11 -04003yrs ago, I got a deskside database server machine, latest 4core @3ghz, 8gbyte memory, a couple tbyte disks, etc. this week, i got a deskside database server machine, latest 4core @3.4ghz (also has hyperthreading), 16gbyte memory, a couple tbyte disks, etc. New machine is less than 1/3rd the price of the old machine.
Was at the dentist yesterday who said he recently had his office PCs (few yrs old, "professional XP") looked at and they told him that PCs had become disposable technology ... less expensive to replace all the old machines rather than try and do upgrade.
--
virtualization experience starting Jan1968, online at home since Mar1970
From: lynn@garlic.com (Lynn Wheeler) Date: 06 May, 2011 Subject: CFTC Limits on Commodity Speculation May Wait Until Early 2012 Blog: Financial Crime Risk, Fraud and SecurityCFTC Limits on Commodity Speculation May Wait Until Early 2012
from above:
The Commodity Futures Trading Commission may not complete limits on
commodity speculation until the first quarter of next year, according
to a filing on the agency;s website.
... snip ...
"Griftopia" covered that the previous spike in oil prices over $100 was result of "19" secret letters allowing entities w/o substantial positions to play. Previously there was rule that excluded entities w/o substantial positions because (their speculation) resulted in wild, irrational price swings. Today, TV business news had quite a bit on commodity speculation wasn't something for individuals to play in because of the wild, irrational price swings. There is temptation to draw analogy with pump&dump tactics.
misc. past posts/reference to griftopia &/or secret letters
https://www.garlic.com/~lynn/2010o.html#59 They always think we don't understand
https://www.garlic.com/~lynn/2010p.html#6 What banking is. (Essential for predicting the end of finance as we know it.)
https://www.garlic.com/~lynn/2010p.html#7 What banking is. (Essential for predicting the end of finance as we know it.)
https://www.garlic.com/~lynn/2010p.html#50 TCM's Moguls documentary series
https://www.garlic.com/~lynn/2010p.html#57 TCM's Moguls documentary series
https://www.garlic.com/~lynn/2011.html#53 What do you think about fraud prevention in the governments?
https://www.garlic.com/~lynn/2011.html#55 America's Defense Meltdown
https://www.garlic.com/~lynn/2011b.html#59 Productivity And Bubbles
https://www.garlic.com/~lynn/2011d.html#21 The first personal computer (PC)
--
virtualization experience starting Jan1968, online at home since Mar1970