You’ll notice there are some companies from Costa Rica & Curacao there, I added in those countries after there were complaints that there weren’t enough providers to be found from BTL to GUY. I called cheese on that, but still opened up to a few more countries not typically associated with the Caribbean.
Doing this exercise brought to mind a need to create something of a Caribbean Cloud Registry. As I wrote that, the word “Foundry” came to mind, but the guys at pivotal sort of have that locked down, I think. But the point remains, a place where you can at-a-glance see who the cloud players are and what they offer.
One of the things I wanted from those doing the evaluations was a discerning eye on whether the service providers in question were offering actual cloud services they built or if they were cloud journey enablers. As an enabler, they may have been either reselling or consulting and supporting an organization’s journey to the cloud.
Cloud enablers providing a necessary service, leading the interested to get to cloud. Without them, organizations can easily get lost in the details of implementation too soon, or make costly missteps when starting out. However, this assignment is not about them. It’s about those who are practitioners in the world of building cloud solutions. This assignment largely involves finding and understanding the context of the cloud service builders.
Not every -as-a-service marketing brochure means the seller is a cloud provider.
Just completed the 2017 edition of the University of the West Indies’ International Half Marathon.
This was a brutal run in beautiful weather. The “race” starts at 5:30 am. Last week, stormy weather was on the cards. Some might believe running in the rain is a joy, but not for 13.1 miles. Thankfully, it was very cool for the duration of my run, no sun, no rain & a light breeze.
I often start off too quickly. The PhillyNet Half Marathon training started just before the middle of the year in earnest and through a series of long and short runs, strength work and other training, the team and I got in to racing shape. 11 of us ran the race, but more than double that at times trained with the team. We had many supporters, on the course and online. But anyway, I still start off too quickly. I may have some overall race pace that I want to maintain, and invariably, my first mile is off. That’s where Endomondo comes in. I start it up on my phone, and at the end of mile 1, the Endomondo Lady announces, “Aye! Stop! You’re going too fast, slow down!”. OK, no, she doesn’t do that. Instead, she announces that I’m running at 8-something a mile or some such and I know,
“Oh, if I want to finish this race in a decent overall time, I have to slow down”. Today, E. Lady was quiet. I got to mile 1 and heard nothing.
Immediately disappointed, I expected it was because even though I started it at the start of the race, bad GPS or some android-y thing had thrown things off. Since a mile had passed, I was not interested in getting it going if it was off. Anyway, I felt good. I just didn’t know what time it was.
I would have run the entire race time-blind, if not for this year’s UWI half innovation: race pacers. Either they were provided by UWI or some other club, but they were golden. Well, not literally. I saw a 1:50:00 pacer and a 2:00:00 one. Keeping up with them would essentially give you a race result of their stated times. The 1:50:00 one flew past me after mile 2. And since I hadn’t seen the 2:00:00 for the first half, I judged to be between those finishing times. But having a pacer in no wise means having the power to finish at those times. So, I had to be careful.
With no E. Lady in my ear, improvisation became necessary. Since there were so many runners, finding and keeping pace with someone on the course became a fun, mini-game. At first there was the lady from +One a Week, I kept up with her for about half a mile. Two ladies overtook me and ran on that spot I started to like – the exact middle of the road on the white line. So, I overtook them to reclaim my land. That cat and mouse lasted for a mile and then they bested me.
Coming up to mile 4, there was loud music. I liked it, then it got too loud, then I liked it again. The DJ seemed very proud to remix “Full Extreme” into “We Runnin Still”.
Could fall down,
We running still… we runnin still.
Could fall down,
We running still… we runnin still…”
I mean, it sorta worked. Something about the loud music added some energy to my run that I intended to use later.
I met up with someone from TT Road Runners and that partnership lasted probably two miles. We got up to the turn together. Mind you, these aren’t announced things. It’s just that as a runner, you realize someone is next to you, they’re keeping pace and either you increase yours to leave them in your dust or you chill and maybe see how far it goes.
She eventually outran me and I was alone. The 2:00:00 pacer passed me at mile 7 or so, but then, my family, immediate and extended was there at mile 8, being raucous and extremely encouraging. Sophie, my 10-month-old, even deigned to give me a look and a small smile. All this probably being a bit much for her thankyouverymuch. I knew the pressure miles for me would be between mile 10 and 11 and a little after. By the time I got past the Eddie Hart ground, which is about a quarter mile after the 10 mile mark, I got into my head that I needed something to distract me. So, I started to count. I just picked a number and counted up to it.
Somewhere between 300 and 400 “counts” I met up TT Road Runner lady. She seemed to be having a hard time. So, I tried to return the favor. Shouting out encouragement to other runners gives me a slight jolt myself. I yelled things like “keep it up”, “this is the hardest one” and such. I don’t know if it worked. Running is a physical thing. Not a talking thing. To a guy who passed me that I was meeting up to, I yelled some more. For some reason, I was legit feeling OK, coming in to mile 11. There’s a hill up to mile 11 and another one up to mile 12.
To keep pace up the mile 12 hill, I pulled out and old trick: Father Abraham. I just sang the song and sort of performed the steps. Way, it’s a long song. It worked, though. All the while, I kept 2:00:00 in sight. By mile 12, I returned to counting. This time, down from 900. 900 because it would take me about 10 minutes to get to 1. I saw my wife, Sophie and #TeamPhillyNetSupporters again. Again providing a burst of joy.
I settled in my mind, a few weeks before the race that the half marathon is a 12.75 mile jog and a quarter mile race. I only really care to finish ahead of people at the very last piece of the run. So that’s what I did. I essentially pelted through the straightaway and for my effort, copped a really good time.
Going in to the run, I was expecting and said as much to do the race in 2:10:00 or 2:15:00. This result was very surprising and I think the combination of decent training with PhillyNet and random race partnerships helped me discover some new things about running.
PS: Oh, as it turns out, Endomondo was fine. I had my audio settings too low, so she was talking, just at a whisper.
I wanted to create a few storage accounts for students in my class to complete an assignment featuring Event Sourcing and Material Views.
So, here’s what I did.
Download/install the latest azure command line interface (cli).
(While doing this, I realized I could have just used the cloud shell. I soldiered on with the dl)
Create a resource group to contain the accounts we’d need.
Create the accounts and output the storage account keys
The command to make a single storage account is pretty straightforward:
But I wanted to also output the keys and display them on a single line. The command to get the keys after the account is created is this:
So, I used the jq program in bash to parse the json result and display both keys on a line. Thus, I created a script that would create the accounts and then output their storage account keys.
This is the script that produced the accounts and keys:
Overall, the longest part of the exercise was dealing with the way the files were being saved in windows vs how they were being saved and read by bash. But the accounts were created and class can get on with assignment 2.
Our text today came from the Cloud Architecture Patterns from Microsoft. This was our first class on cloud architecture and so we spent time looking at several popular patterns, including,
I called back to that night I built a website for tracking the St. Joseph By-election in 2015 and how I could have used this on the client side when things became too intense for the database server.
The writing on this is great, but Clemens Vasters explanation of Sagas really brought this home to the class.
Competing Consumers Pattern
This is one of the patterns we’ve used a lot for our messaging solutions at Teleios, so it was easy to walk through and talk about messaging demos featuring it with the class.
Compute Resource Consolidation Pattern
I used this pattern to start the discussion on how containers are a great way of consolidating a set of compute-based activities. From the patterns guide, there was a great example though of the need for care to ensure that the right kinds of tasks are consolidated. Tasks with dissimilar characteristics can reduce the efficiency gained from pursuing consolidation.
In talking about this pattern, this amazing quote was uttered in class:
The #serverless paradigm is the Mr. Meeseeks of technology. Don't be a Jerry. 🙂 Said this in Cloud Technologies class today.
The next assignment is based on Event Sourcing and Materialized Views, where students have to come up with a scenario that features the use of ES and MV and then build a demo that explores their chosen scenario.
It was also cool to see the call back from our big data class last week because I highlighted that Apache’s HBase might be a good data store for building a solution involving Event Sourcing. There’s also a great article talking about the Command & Query Responsibility Separation pattern in the context of Event Sourcing, too.
In Cloud Technologies class today, we used both the course outline and the notes from MSFTImagine’s Github repo to talk through the differences in service offering.
I used the canonical service model responsibility chart to start the conversation off.
It’s fairly straightforward to talk to these divisions, of course. I often use it to drive home the NIST 2011 definition of cloud services. With emphasis on the service delivery models.
In today’s presentation, one of the things that jumped out at me was the slide that provided a distinction between SaaS Cloud Storage and IaaS.
Finally, when talking about the ever versatile Salesforce, and how its PaaS solution works out it reminded me of the Online Accommodation Student Information System (OASIS 🙂 ) that I had built when I was in undergrad.
I’d built OASIS as a commission for the Office of Student Advisory Services. It was a tool to help off-campus students more easily find accommodation. Prior to OASIS all the information was a notebook in an office. It was built before I learnt about the utility-based computing of cloud. I’m thinking about using that as the basis of an exploration of the architectural changes need to move an old service to the cloud.
Hopefully, I’ll be able to revisit it when we touch on Cloud Design Patterns.
Started back with the UWI Cloud Technologies course today. This class was an Introduction to Cloud generally, with some conversation about the course outline and expectations for assignments.
We still in the process of confirming the course outline, so I’ll share that next week. But I used the slides from the technical resources provided by the Azure Computer Science module on cloud technology.
On my way to class I met up with Naresh who runs the UWI’s Department of Computing and Information Technology servers. He gave me a quick tour of their deployment. I’m looking forward to him sharing some stories from setting up that environment in our IaaS classes in a few weeks.
Just a small note to say I’ve added in to Nurse Carter the ability to provide information about shelters and emergency contact numbers.
Of course, there’s a wider story around this. About #OpenData and being able to leverage that for future needs. But I won’t go in to that now.
I did some scrounging around online, found the PDFs and stored them on github in case the Office of Disaster Preparedness and Management’s site goes down during build up to a hurricane again. And that’s about it.