Of the numbered choices below, which color combination was used as the “read-only” data entry field in EpicCare Ambulatory/Legacy?
The background color of most of the UI was “battleship gray” as shown above.
You may remember from my post The App is Too Fast!, that Visual Basic 3 (and Windows 3.11) had a very limited color palette. Even as I created that sample application in Visual Basic 3 for this post, I was again surprised how few colors we had to choose from!
In any case, I’m waiting for your answer.
Come on. Stop trying to skim ahead to see the answer before you pick.
It’s definitely not #14. As the background color is a dithered orange, there wasn’t a reasonable color combination that worked well for showing text on it correctly (as you can see from the screen shot).
There were very few (if any) design guidelines for how a graphical user interface should LOOK back then.
OK, alright. It was #3. A yellow background and black text.
I never heard a concrete and well-reasoned explanation for the choice of yellow. What this meant was that if there were fields/inputs on a form that were not currently available due to configuration or selections made by the user, instead of a typical white background, the input background would be yellow.
On a screen like EpicCare Ambulatory Order Entry back then, where 95% of all the fields that might be used by any combination of an order (labs, prescriptions, tests) were placed, many of the fields would be shown as yellow at any given time.
”White means work.” —Nurse from Kaiser Permanente NW
During a few day site visit to Kaiser Permanente NW (many years before the big sale to Kaiser Permanente National), I and a few other developers were following a nurse around during her shift. She had just walked up to a shared terminal and started to place an order. That’s when she said that.
”White means work.”
What was interesting about that was that she had just landed on one of the most complex screens in all of EpicCare Ambulatory at the time. It was a chaotic orchestration of labels and inputs and list boxes, and tables and … literally everything seemed like it was on this form. Imagine the drawer we all seem to have, the one with every miscellaneous thing in it to handle nearly anything that life throws at us.
That drawer was Order Entry back then. If you could find it all, you could finish the task.
As she moved through the necessary selections of the particular order she was placing, the seemingly endless fields on the screen changed to “white” and that’s when she uttered that amazingly memorable phrase.
She was disappointed by how much of the screen was going to require active input and energy from her. When we asked for specifics, she mentioned how the “old way” was faster because there was far less tedious fill-in-the-blank to make the computer happy.
Her complaint wasn’t about the yellow — it was about the fact that this modern system frustrated her and signaled that she’d need to do extra work to satisfy the computer.
Is the work that the customer is doing for your software or for them?
Think about that as you create your next app or workflow. Please.
The Yellow Rectangle Road …
During our initial Cadence GUI design sessions, we avoided the “yellow” backgrounds for read-only fields. While it was a departure from the established style of EpicCare, the users of these applications would have zero cross-over at that time. We picked #10. There was some internal moaning, but we held fast.
”Hold the grays!”
I know there were limited color choices, but for me and the way I’m wired, bright yellow rectangles drew my attention unnecessarily when these should have been no more prominent than the white text input fields. In fact, I see them first as I suspect is the case with many others.
The yellow background remained part of EpicCare for more than a few years. Thankfully, it didn’t continue through the EpicDesktop/Hyperspace transition.
What color would you have picked (given the choices above)?
If you want to compile a Tree Sitter grammar to WASM, you might find this Docker Build/BuildKit example helpful. I didn’t want to set up all these tools directly on my machine, so this provides a helpful one long step to isolating the build and tools, and at the end, saves the created WASM file in the directory from where you launched the build (you can change the location by adjusting the output argument).
You can use build arguments or change the file directly.
Replace the TREE_SITTER_GRAMMAR_GIT_URL with the git URL where the grammar is located. For example, it might be:
Replace the TREE_SITTER_NAME with the destination directory name that is used for the clone. For example using the repository name from above, it would be tree-sitter-sqlite.
Then, from the folder where you’ve saved the Dockerfile:
I didn’t spend too long on Cadence before switching to another project. But I have one short and very memorable moment from the Epic User Group Meeting in 1995 while still on the Cadence team. Cadence GUI was to be demonstrated for the first time in front of all attendees. We were sitting in a very large room at the Alliant Energy Center complex as the attendance had outgrown the Concourse Hotel space (bigger, but not better).
Three of us sat in the left front row waiting for the main session to start. (It was an awful place to sit so we weren’t taking up seats that were desirable by customers). We were mostly just joking around with a bit of tense energy.
Carl was “headlining” though and had to do a presentation before we did our debut demo-dance. He sat down next to us, visibly nervous. He was normally composed, so it seemed out of character from my experiences with him. I don’t remember the exact words, but this captures the moment:
I have about 10 minutes worth of content. I’ve practiced and I’m going to do it in about 2 minutes. —Carl
We laughed thinking he was joking, but he assured us, he was not. As ever helpful coworkers, we provided no great advice to calm his nerves, as we had none. It was too late anyway.
Bless his heart. He got up in front of the audience and slowed down and went over his prediction by no more than a minute or so. It was RAPID FIRE INFORMATION. It was awesome and hilarious: it was impossible to keep up. 🤯
It took me more than a few years to be far less nervous doing my own presentations.
It’s funny that he gave me advice after a sales presentation … “you know this stuff inside and out. Just talk about it.” And I realized, I did! I didn’t need to write everything down. I knew what I wanted to talk about. If I left something out, I was the only one that knew really. If I ad-libbed a bit of content, again, I was the only one that knew.
While “knowing the material” isn’t enough to get rid of presentation nerves for a lot of people, that little piece of advice helped me a lot. I started to remove a ton of words from my slides after that. If the words were there for me as anything more than a direction/general topic, then I’d delete them. The words and images/diagrams were there for everyone else. Not me. That was a shift for my approach, but it worked so well.
If I didn’t know the material, then I learned it. It shouldn’t be a surprise how much better a presentation can be when the speaker is confident in the material. It doesn’t make them a great speaker, but it certainly helps establish credibility which is far too often lacking in technical presentations.
Additionally, I would think about what questions I’d ask if I were listening to the presentation and prep for those specifically (and modify content accordingly). If someone caught me off-guard with a question that I’d not considered, I’d write it down and not let it bother me that I’d missed something.
”I don’t know, but I’ll find out” is much better than trying to sidestep and faking it.
The Cadence GUI Demo
It went fine as demos go. I recall we had a glitch, but recovered (and we made it clear it was a work-in-progress). Was there thunderous applause after the demo and a standing ovation? Abso-gui-lutely! OK, not so much. A polite applause. (“Clap clap, when is lunch?“) Since it wasn’t a new product, the interest was mostly about new workflows and improvements to scheduling efficiencies that would result.
Presentations
What’s helped you doing presentations? Or what still bothers you about doing a presentation in an unfamiliar setting? I’ve heard so so so many people over the years say: “I hate doing presentations.” Are you one of them? If you need help with your corporate review or “big pitch” presentation, please consider my 1 or 6 hour consulting services.
I really appreciate you stopping by and reading my blog!
You might not know that each Epic blog post takes me several hours to write and edit.
If you could help me by using my Amazon affiliate links, it would further encourage
me to write these stories for you (and help justify the time spent). As always, the links don't add cost to the purchase
you're making, I'll just get a little something from Amazon as a thanks.
I'll occasionally write a blog post with a recommendation and I've also added a page dedicated to some of my more well-liked things. While you can buy something I've recommended, you can also just
jump to
Amazon and make a purchase. Thanks again!
If you buy something from a link, Acorn Talk may earn a commission. See my Affiliate Programs statement.
This is embarrassing. We’ve been using a combination of a few simple devices as a very reliable way to water some trees we planted last summer, and today I decided I’d provide a few details about the products. But, when I went to Amazon to grab a link for the primary part of the solution, it’s no longer for sale! On it’s own, that’s not too surprising as resellers come and go seemingly every hour on Amazon. This one however is because the small company (person) making them has retired and has stopped making them. 😒
Over the years, we’ve tried a lot of different ways to water new trees and bushes. So many failed miserably. From those large and small tree watering bags (thanks for the leaks!) to just the plain hose end. The key to watering trees is generally go slow — so that water can saturate the roots.
The breakthrough for us in convenience and reliability was the discovery (and subsequent purchase) of this: Waterhoop.
It’s not complex, but that’s what has worked so well for us. Once connected to a common hose and the spigot is turned on, you control the water volume directly on the hoop (so no walking back and forth to the spigot to get it “just right”). Water drains or sprays (depending on volume) through a series of small holes around the hoop ends. And, that’s it. As it’s flexible, it adjusts to a variety of situations. Having local control is brilliant — it’s such a simple feature that saves time!
As we live in a suburban neighborhood now, the water pressure changes during the day. Before we had this combination, we’d set up a hose at the base of a newly planted tree, set a timer for an hour, get the flow just right … and return an hour later to find that the water was just dribbling out, nothing like we’d originally set it. Now, with the simple flow adjuster at the end, we can be assured that the water amount we want will be the amount an hour later (as we turn the flow up at the spigot far beyond what we actually need as it’s now regulated near the target rather than the source).
While setting a timer on a phone, smartwatch, etc., works, it’s been even more useful to put a mechanical timer at the end of the hose, again right near the target. So, we can set it for 60 minutes, and not be concerned that we’ll over-water if we forget to go immediately out when a timer signals. We’ve had digital timers on the source end, and frankly, they’re more than we needed for this, and not needing a battery has been great too!
The Orbit Mechanical Watering Hose Timer has performed flawlessly for us. By moving it to the end of the hose, we can quickly set the timer, adjust flow as needed (with the flow adjustment on the Waterhoop), and be done with it.
While these things may seem unnecessary (and truly are), they’ve been a big time saver for us as we were watering weekly, and more importantly, we were able to control the amount of water far more precisely than before.
Overall, I’d totally recommend both of these if you’re watering trees, bushes, etc.
OK, but I know — the Waterhoop isn’t available anymore. I’d picked the Waterhoop because of the very good reviews and the fact that it was made in the USA. Instead, if I couldn’t make one myself (I’d try!), I’d switch to using a soaker hose for trees with a few tweaks. I’d either get rid of the “Y” part or make it a quick connect on one end with a product like the Eden Quick Connect. That way, it would be a snap to remove it from one tree and move it to another. If I wasn’t concerned about a “true” loop, then I’d use a hose end cap. I usually have a few of those around. It’s possible that the drip of these would be slower than I’d need for our soil, so I’d carefully make a series of very small punctures around the hose to increase the drip speed. It probably wouldn’t matter much, but rather than having two potentially fiddly valves on the “Y” adapter, I’d add an independent coupler with flow control.
I really appreciate you stopping by and reading my blog!
You might not know that each Epic blog post takes me several hours to write and edit.
If you could help me by using my Amazon affiliate links, it would further encourage
me to write these stories for you (and help justify the time spent). As always, the links don't add cost to the purchase
you're making, I'll just get a little something from Amazon as a thanks.
I'll occasionally write a blog post with a recommendation and I've also added a page dedicated to some of my more well-liked things. While you can buy something I've recommended, you can also just
jump to
Amazon and make a purchase. Thanks again!
When Cadence GUI entered the Epic stage, the team was provided with a complete copy of Legacy/EpicCare to do with as we saw fit. There was zero process in place for any formal code sharing at the time, so we stripped our copy down to the bare bones leaving only a small communication engine and a developer hub Visual Basic Form (that I plan on talking about later).
Options for input and output with a socket from the MUMPS server were very limited back when EpicCare was started. While there were built-in functions for communicating to external devices, they were focused on brief communications and had very limited control options. Acting as a “server” for a long running persistent communication channel was far more challenging. Further, the support across vendors for options wasn’t consistent.
Faced with this communication challenge, the Legacy team built a resourceful alternative. All MUMPS server hosts supported connections via a (now thankfully waning) protocol called Telnet. The Telnet protocol shows its age these days and isn’t commonly used (and was phased out at Epic decades ago). But in 1992, it was a common service that was available on all operating system platforms Epic and customers were using.
As a brief aside, as the Telnet protocol was an operating system service, each OS (and version!) had its fair share of quirks. It was fortunately uncommon, but we would encounter situations where a version of an OS, for example HP/UX, would improperly handle a documented Telnet command. Of course, as bugs often go, it was only in certain circumstances and combinations which made troubleshooting cumbersome.
MUMPS_Command was Born
The Legacy team exposed the functionality in Visual Basic by creating a function called MUMPS_Command. When the Visual Basic application launched, using a configuration file that was dubiously secured, tied with an application hard-coded decoding key, the host application would make a Telnet connection to the MUMPS server host OS.
It would then send … locally stored credentials … (sigh, yes, that’s how it worked back then). Upon a successful login and some scripting magic in the OS, a MUMPS job (process) was started immediately and began to execute a specific MUMPS routine. This was a “captive” session.
If someone knew the user name and password for this special MUMPS_Command user, they’d be launched directly into the Epic created protocol for communication from client to server.
Some of you may have used a Telnet client to connect to a OS service for troubleshooting.
BASH
> TELNET localhost 5555GET / HTTP/1.0
It definitely doesn’t put the “fun” back in functional though.
After a brief exchange of a control sequence to verify the connection and some important settings, MUMPS_Command was ready for duty. Although “doody”(💩) may be more appropriate for the early versions.
How did it work?
I cannot say that the early versions of MUMPS_Command were robust, or secure, or reliable. They weren’t. Telnet isn’t secure. The protocol and services are terrible and have no baked-in security. And yet at the time, Telnet was used most often for connecting to a remote host. Using Telnet back in the mid 1990s wasn’t unusual, so Epic using the protocol as a connection didn’t raise any general concern across customers. (Live encryption of this data on the wire would have been unheard of back then given it was running on a secured Intranet and the massive increase in computing power required made it a non-starter).
The core idea was that a captive MUMPS job/process was either running MUMPS code or waiting for input from the end user via a terminal or pseudo-terminal (teletype TTY or pseudo-teletype PTY). Given that the Visual Basic client (Cadence and EpicCare) were remote connections to this captive process, the MUMPS code was in an infinite loop.
MUMPS
COMM NEW input,done,TIMEOUT SET TIMEOUT=300WAIT ; READ input:TIMEOUT GOTO WAIT
It’s a slight oversimplification of what the code looked like, but it’s not far off. It’s not done yet. (I also used non-abbreviated MUMPS instruction names).
The Visual Basic client, over the Telnet protocol would “send” a request to the captive MUMPS session by “typing” it and sending a carriage return (0x0D). The connected MUMPS job would receive the text and store it in the variable input as shown above. Using the syntax as shown above, the MUMPS READ command only ends when it receives a line of input (ended by a ‘terminator’ character which included the carriage return).
Now that the input has the request sent from the Visual Basic application, what’s next? The team decided on what became an unfortunate choice a number of years later. It was effective and easy. But it was fragile. And it offered no reasonable security (especially in the early versions).
MUMPS has a wonderful command/instruction called Xecute (think of it as eXecute). The string provided to this command can be any valid MUMPS expression is immediately executed. Many interpreted languages have a similar feature. JavaScript has eval for example. eval has been used many times over the years for perfectly fine JavaScript browser code, humorous hacks, and too many noteworthy nefarious reasons.
Remember, this Xecute command allows execution of any valid expression.
Some of you may be shuddering already. Good!
MUMPS
COMM NEW input,done,TIMEOUT SET TIMEOUT=300WAIT ; READ input:TIMEOUT IF input="**END**" GOTO EXIT XECUTE input GOTO WAITEXIT ; WRITE !,"**ENDED**"
(It’s weird for me to type out the full command XECUTE, as I don’t know that in my 25 years of Epic I ever used anything but the abbreviated X!)
The loop reads the input, checks to see if the client connection wants to end the connection, and if not, executes the expression that was passed.
As requests need a response, the expression sent needed to store the result in the variable X.
MUMPS
S X=$$getFut^SCHED1(" 20240705",53711)
The Xecute command parsed the expression and executed it. In the case above, the code is calling into a second MUMPS routine to get future appointments for a specific patient identifier from a specific date.
Upon evaluating the expression, the MUMPS_CommandWAIT loop would write to the Telnet session whatever value was stored in the variable X.
MUMPS
COMM NEW input,done,TIMEOUT SET TIMEOUT=300WAIT ; READ input:TIMEOUT IF input="**END**" GOTO EXIT XECUTE input W X ; sent back to client GOTO WAITEXIT ; WRITE !,"**ENDED**"
Annoyances
MUMPS for much of its existence has had extremely short string storage/length limits in memory and when stored in MUMPS Globals. Because of this limitation, and because of how the Epic invented protocol worked at the time, it meant that the Visual Basic developer had to make certain that the MUMPS_Command requests did not exceed the string limit (this number was passed to the client as part of the initial handshake).
Long requests and responses had to be broken into segments. It was very common to need to send a “request # of N” as part of a series of MUMPS_Command calls. A free text field on the client that was longer than allowed by MUMPS required that the developer break the string into pieces (pun intended) and send them in chunks. Reverse that for the server sending that same field to the client. Learning to return a “is there more work to do” was a common pattern on client and server.
Line terminator characters had be carefully transformed before being sent over Telnet to the server. If an errant 0x0D (carriage return) was in the data, it would cause the XECUTE to begin evaluation of the expression immediately. Then … the whole communication channel was broken as the client and the server protocol would be mismatched — the protocol was a simple state machine on both ends. Either waiting or receiving. There was no way for an out of band communication to be handled properly.
Line feeds and carriage returns were transformed (or $TRanslated in MUMPS) to non-line-terminating characters 0x01 and 0x02. Occasionally, those would be the source of weird Telnet service issues over the years. 😒
Little Bobby Tables
If you’ve done any database development in the last decade, you’ve likely seen this:
Well, MUMPS_Command had a problem that was an extension of the XECUTE command. The XECUTE command evaluated and executed all valid code.
If the client sent an unsanitized string, like a patient name with a value of: JORDAN, MICHAEL ") K ^ER S %1=(""
This would get sent to the server:
MUMPS
S X=setName^DEMOG("JORDAN, MICHAEL") K ^GLOBAL S %1=("")
Rather than the intended:
MUMPS
S X=setName^DEMOG("JORDAN, MICHAEL")
Needless to say, there was some non-Midwest nice words used in the offices for a few days as developers scrambled to fix the issue.
It really wasn’t ideal that a user of a client application could say, delete an entire global structure: K ^GLOBAL. That’s syntax in MUMPS for deleting an entire GLOBAL (KILL ^GLOBAL). This would be similar to a SQL statement where an entire table is dropped (and a huge disaster).
I’m not aware of any customer related security issue to this design. The lack of properly handling quotes caused application stability issues most noticeably.
There became a little “quote” coding challenge to verify that there wasn’t unexpected quotes and that all Xecutes worked with expressions that had been sanitized properly.
Licensing
Epic was more frugal in the 1990s than it has been in decades. When the company needed to spend money, it did, but reluctantly, and not without putting up a licensing fee fight. For many years, Epic relied on a third party vendor for providing a Visual Basic compatible Telnet client (weird weird since Epic has so often been considered a “We Build It Here” software factory).
While I believe there were two vendors that had been used, the most used vendor was from a company called Distinct. The product was called … Distinct Telnet. (HOLY MIDWEST COWS READER! Distinct exists and still sells an ActiveX version via what seems to be their FrontPage 98 web front end using ASP web pages). Frankly, it made little financial sense for Epic to write that code. Visual Basic wasn’t fast enough yet to make it viable (as it wasn’t compiled to native machine code in early versions), and there was no one with experience in creating a Visual Basic 2/3 component using C/C++ for a communication protocol, and with Telnet.
Eventually Epic reached a licensing deal with the authors of the Telnet component that seemed high at the time. The real nuisance was that it was a per-install license for customers and Epic internal use rather than a site-license. Epic software wasn’t licensed “per-machine” so tracking client devices was unnatural.
Hello Epic Telnet!
Years later, when I was the team lead of Foundations, we embarked on creating our own Telnet implementation as the costs of the Distinct Telnet component had risen and it was difficult to justify the cost to all customers — and more importantly it wasn’t as fast as we thought it should be. During the course of a few months, a talented developer created “Epic Telnet” and we tested it internally (albeit slowly and with a significant amount of internal trepidation); after a significant amount of internal testing we began to cautiously install and use it at new customers. While my team and my own feelings were more optimistic and pragmatic, I know that others felt this new Epic Telnet was far too risky to implement and far too outside of the norm for us to be developing. While I understood their apprehensive and cautious response, the deployments went well and we demonstrated how the new communication system was more reliable and faster than it had been using just Distinct Telnet (in part, not due to a fault in Distinct Telnet, but more being able to tune our code to our exact needs). The addition of Epic Telnet removed a cost and several pain-points related to the older component. Strangely, I know it took YEARS for the older Telnet functionality to be fully removed and not used. It was a slow process, not entirely atypical for healthcare IT I suppose across the spectrum of customers and cultures.
A Positive Connection
A huge amount of work was performed over the years without issue with this simple protocol created at Epic. Sometimes simple to start is all that is needed, and for Epic, it was a great early fit. It had growing pains over the years though.
We did some other cool things over the years with this Telnet based communication system, but this post has gone long …, so for another time!
In the mid-1990s, using the name “MUMPS” was becoming less tolerated by potential customers and the ISO standard had approved using M as an alternative. For reasons that are still amusing to me, Carl “strongly requested” that all uses of MUMPS_Command be renamed to M_Command. While customers weren’t encountering the name (unless they were browsing the source code, or doing their own custom development), dropping the name meant it was less commonly spoken and less used during every-day conversation at Epic. M was the new better name, for reasons. It later became RPC_Command fully dropping the M in favor of Remote Procedure Calls as that sounded more inline with terminology and techniques that were gaining favor at the time (mid-late 1990s).
I’m going to stick with MUMPS for a while longer though, unless Carl pays me to stop. 🤓
I really appreciate you stopping by and reading my blog!
You might not know that each Epic blog post takes me several hours to write and edit.
If you could help me by using my Amazon affiliate links, it would further encourage
me to write these stories for you (and help justify the time spent). As always, the links don't add cost to the purchase
you're making, I'll just get a little something from Amazon as a thanks.
I'll occasionally write a blog post with a recommendation and I've also added a page dedicated to some of my more well-liked things. While you can buy something I've recommended, you can also just
jump to
Amazon and make a purchase. Thanks again!