On 2015-04-11 05:22:29 +0000, IanD said:
> The OS needs to be the ubiquitous entity that people and processes
> interact with the way they wish to interact with a system, not be
> straight jacketed into having to deal with system limitations
I'm going to use "interact" slightly differently than you have.
Servers aren't visible to end-users. Or shouldn't be. The herds of
servers are and should be visible only to IT staff. Details of the
servers and the operating systems to the most technical of the support
staff and to the developers. Fewer sites even have technical support
staff and staff developers, and many have outsourced that. IT is not a
competitive advantage for many organizations, to paraphrase a common
comment.
> The presentation layer and the data engine should be separated with the
> OS handling the complexities and optimization behind the scenes of how
> to store the data and retrieve and manipulate the data - this is why MS
> were looking at the DB file store concept. A rehashed ODSx isn't going
> to get us there
In terms of the end-users, the user presentation layer is long gone
from OpenVMS, beyond existing applications and developers, and the
presentation layer is certainly from most servers in general.
In terms of the technical staff, servers are replaceable and
interchangeable units, running stock installs with few or no
customizations. That's not the same world that begat OpenVMS.
Further down the stack, yes, it's increasingly all automated and
self-tuning and with logging and the rest looking at whole racks and
whole datacenters, and not at individual boxes. The most technical
folks and the developers get to see this layer. But again,
self-tuning and mass deployments are not where OpenVMS came from.
> If there is a limitation, fix it but don't have the users of the system
> have to know every performance trick to get adequate performance from a
> system
Alas, you're going to have to know some tricks for your platform and
you're going to have to know your hardware and software, if you're
pushing it hard enough.
But for most folks, yes, self-tuning, self-administering, and
self-overwriting and self-healing — without losing data — when repairs
or replacements are necessary. Simpler.
> MS wanted some time back to move to a DB file system. From what I have
> read lately, it looks like that concept may have been revisited with
> Windows 8 and some of the ground work added for a future move towards
> this idea again
File systems are databases. Always have been. They're just using
human-understandable metadata tags. As the scale of the data involved
and the computers involved increases, the humans and the human-friendly
metadata becomes the weak point of the classic file system designs.
Computers don't care if the blob of data is tagged as "MyCoolProgram.c"
or "375620F9-03B5-42C1-471D-A21B1DA89F73", after all. Computers don't
care if the filenames are sorted into any particular order, either —
unlike your particular case and your ginormous directory, which have
directories and names which are sorted. So how do you best deal with
humans, and are what the more experienced humans expect — classic file
systems and folders or directories — really the best solution for
various problems? Some yes. Some... not.
> If you want to see how far behind the times ODS2 is, have a quick
> glance at the wiki for file storage systems
>
>
http://en.wikipedia.org/wiki/Comparison_of_file_systems
>
> glance at the Limits section as well and you'll see ODS2 is rather
> crippled compared to more modern systems, even ODS5 has noticeable
> limitations
In terms of its design, ODS-5 is ODS-2, for all intents and purposes.
The on-disk differences between the two are FI2DEF versus FI5DEF, and a
constant or two. The filename changes and the directory depth changes
are largely the removal of assumptions and limitations within the
parsers and the applications; in the software. Some related
limitations in other components such as the DCL command line still
lurk. Parts of the ODS-5 design such as the escapements explicitly
worked around the lack of UTF-8 support in DCL and in many parts of the
character-handling support within OpenVMS, for instance.
To me, the newer and layer-breaking file systems are more interesting —
ZFS, etc — as they are new approaches to solving problems, where
OpenVMS uses largely isolated pieces including ODS-2 or -5 and the XQP,
XFC, HBVS and related pieces. Component isolation and layered designs
help in many ways, but can also means it's far more difficult to
leverage knowledge that's available in one layer from within another
layer. Catching up ODS-2 or ODS-5 with what's available now for
isolated file systems is no small effort, and you're still going to end
up five years behind what other folks are doing, and probably ten years
behind ZFS or other more "radical" changes. From over seven years
ago:
<
http://arstechnica.com/gadgets/2008/03/past-present-future-file-systems/>
But again, we don't know what VSI has in mind here for their
replacement file system, either.
> The system I currently support runs and old application. It generates
> 10,000's of files every day, around 2 files a second, all plonked into
> a single directory
> This is NOT abuse by myself or abuse by anyone for that matter (as was
> implied by someone else), this is merely the consequence of dealing
> with an old application that was never designed to scale up to the size
> that it is running at today. It was supposed to be decommissioned years
> ago but these days businesses are not application swapping as much as
> they are application enhancing
Can VSI extract enough revenue to make dealing specifically with your
particular case and that of similar folks viable? Will — as most
sites do — the applications sit unchanged for decades? There's
probably not enough revenue, when you trade off having to maintain the
ability to support and patch those old versions and those old
configurations, and trade off that investment with the benefits that
might accrue from moving forward. Put another way, the old model of
OS and application and hardware support for decades is just not viable
for many of the vendors. Not at the prices that can be charged for the
new software and for support in this era.
Times change.
Servers and software are five years old? Roll in the replacements, and
save on power and support and the overhead with the replacements, and
deal with upgrading your applications as part of moving from one
long-term-support software version to the current long-term-support
version available from your preferred vendor. This approach is
disruptive to expectations, too. It didn't used to be the case that
replacement servers were so much more efficient. This revolving door
model of server deployments is certainly not what the current OpenVMS
folks are accustomed to. Various folks using OpenVMS here have servers
that are ten years old, and sometimes more. AlphaServer DS25 boxes
aren't the bleeding-edge of server technologies. How long has it been
since your ginormous-directory application was substantially upgraded
and/or ported forward?
> How many implementations do you see catering for real future growth and
> real business changes at design time? In an age where companies don't
> even know who/what/where/when the markets that they will be competing
> with tomorrow, it's just implement and hand that other future stuff to
> the operations domain to deal with
Part of that inherently means migrating data and chucking out old
applications, and evolving and updating the surviving applications.
That's painful to existing users, and painful and risky to the vendors.
> People have identified an issue with directory performance in VMS, and
> enough of them to warrant that it's not isolated complaints or people's
> imagination
For your case, try an SSD? I'd also look to inboard flash, but AFAIK
neither OpenVMS nor the Integrity servers have support for that
hardware configuration.
> Splitting up files across directories is a workaround, not a solution.
> Having to do this and having to change reporting scripts/interfaces to
> deal with these workarounds rob people from spending time on productive
> measures - less maintenance on a system, not more should be the key
> driver
You're presuming that there are ways to change certain behaviors of
ODS-2 and ODS-5 behavior, short of wholesale replacement. ODS-2 and
ODS-5 maintain sorted lists of files stored in directories scattered
around the disks. This sorted linear list design fundamentally
doesn't scale. Scattered sorted-list files containing metadata is a
design which involves extra disk seeks for those operations, and thus
there are usually caches specifically for that data, and other
design-derivative details fall out of these decisions. Going to a
red-black tree effectively means the wholesale replacement of the file
system won't make folks happy if/when details change, and that's before
discussions of the entirely understandable aversion to risk that many
people have. There are always trade-offs.
> System automation techniques should be handling these types of system
> limitations automatically so that users and applications are isolated
> from what goes on underneath instead of trying to force the world and
> how it interacts with a system to pussyfoot around what is essentially
> a system limitation
Certainly for a number of cases. For others, this not-knowing approach
verges on installing screws with hammers. It works, but lacks finesse.
> Where I work, systems automation is ramping up. Companies do not want
> to pay for experts to be available at the coal face, they want
> automation robots handling more and more system functions, leaving only
> real issues to be escalated up to L2/L3 support levels, whom the
> organisation don't want anyhow because they are too costly
You're very far from the bleeding edge of this consolidation process.
> I said this before, the VMS system manager must eventually die (I say
> that with the greatest reluctance) because costs are driving high
> maintenance systems out the door (by high maintenance I mean systems
> that require continual tuning, or experts on hand to ensure users play
> nice with them)
As of last summer, OpenVMS was receiving vendor support, with
exceedingly limited enhancements, and with the eventual phase-out of
support basically scheduled.
HP is still executing on that plan, too.
VSI may well extend the long-term usefulness of OpenVMS, and could
reverse the long-term trends for OpenVMS, assuming they can get to
stable revenues.
As for the automation and integration that's necessary here, yes.
That's also no small project for VSI.
> How to fix directory performance on VMS?I don't think it can be or
> should I say, I don't think it can be fixed to scale to the demands of
> enterprise sized systems
VSI has mentioned plans for a replacement file system, but no details.
> The internet of things is coming,
Already here. q.v. CVE-2015-2247
> the OS's that get a lions share of dealing with the data spewing out
> from your toaster and wearable technologies will go to those systems
> that can adequately deal with huge volumes of small data snippets. VMS
> isn't ready and pandering to ODSx/RMS to heal itself will not work
> either
Whether OpenVMS can be extended and enhanced far enough and fast enough
and profitably enough remains to be determined. There's presently
certainly a niche for OpenVMS among existing users. Going beyond that
installed base involves years and decades of work, very large amounts
of cash, accruing the revenues to support the effort, and all this in
an environment where the already-competitive products are also moving
forward. It's a project that's vastly larger than your 10,0000 file
directory problem.
> Better to allow VMS the ability to support multiple file systems than
> work with something that has limitations that were set when the world
> was a very different place
> Create the new, transition across what is still relevant today to me is
> a sensible approach
That request has been around for a while. Replacement file systems are
difficult and disruptive projects. Replacement file systems are also
slowly adopted at many sites. Around decade ago, OpenVMS hacked out
the last of the disk-geometry-sensitive I/O code, for instance. That's
before any discussions of Spiralog, too.
> VSI I suspect will be busy enough just getting larger storage devices
> working on VMS, they cannot do everything as they are catching up on
> years of HP neglect
The 32-bit block number is all over the place within OpenVMS, and in
various applications. Much like the move from 512-byte memory pages
to 8192-byte memory pages, any move from 512-byte blocks and 32-bit
addresses to 4096-byte Advanced Format Disks and 64-bit addresses won't
be painless. Then there's that storage is moving off the I/O buses and
in-board, which means that dealing with the old "restart-reboot"
battery-backed RAM logic from the ancient OpenVMS consoles and ancient
OpenVMS-related hardware is coming back to the forefront in other
environments — it's already common in some computing environments.
Make no mistake: VSI has a huge amount of work ahead of them, and their
biggest goal is undoubtedly to get their revenues going, and to get
their profits trending in the right direction. For the folks that are
dependent on OpenVMS, this situation also means making a decision to
continue with HP support and HP's roadmap for OpenVMS, or paying VSI
for software that — unless you need those Poulson servers — won't be
all that immediately beneficial to you. Possibly in addition to paying
HP too, depending on local requirements. That sort of application
introspection and that decision process is not going to be an easy
decision for many entities, either. Also whether OpenVMS — whether at
HP or at VSI — will continue to meet your needs for your existing
applications, and also whether OpenVMS will be useful for wholly new
development, and what changes might be necessary or expected or
desirable.