Ken Rimple

GraphQL, the new Contender to REST

The Representative State Transfer (REST) protocol has been the king of remote access protocols for web applications for well over a decade. The general pattern: expose “nouns” (Customers, Activities, Employees, Tasks, Sasquatches) as URLs (/api/sasquatch/32) and access them via HTTP “verbs” such as “GET”, “POST” (create), “PUT” (update), or “DELETE” (umm, well…). The content type … Read More

Amazon Workspaces – Desktops in the Cloud

I’ve been experimenting with Workspaces for a week now, and have configured an instance for aws re:Invent for those times I want to hack. Coupled with an iPad Air, a bluetooth mouse, and the keyboard case, I have desktop I can pull up anywhere.

Connecting to AWS IoT Core – IoT on AWS – A Philly Cloud Computing Event

AWS IoT provides connectivity to IoT devices through HTTP and MQTT. In this session we learn how to leverage AWS Core IoT as an MQTT broker, how to connect your devices using a client certificate, how policies can enforce data security, and how rules are used to move data elsewhere in the AWS infrastructure.

Coping with Aging (Data) – IoT on AWS – A Philly Cloud Computing Event

Data has different purposes over time: when fresh, it can be used for real-time decision-making; as it ages, it becomes useful for analytics; eventually, it becomes a record, useful or perhaps not. Each of these stages requires a different approach to storage and management, and this talk looks at appropriate ways to work with your data at the different stages of its life.

That’s not a Data Lake, THIS is a Data Lake – IoT on AWS – A Philly Cloud Computing Event

This talk will review two common use cases for the use of captured metric data: 1) Real-time analysis, visualization, and quality assurance, and 2) Ad-hoc analysis. The most common open source streaming options will be mentioned, however this talk be concerned with Apache Flink specifically. A brief discussion of Apache Beam will also be included in the context of the larger discussion of a unified data processing model.