The last time I spoke to Red Hat CEO Jim Whitehurst in June 2016, he had set a pretty audacious goal for his company to achieve $5 billion in revenue. At the time, that seemed a bit far-fetched. After all, his company had just become the first open source company to surpass $2 billion in revenue. Getting to five represented a significant challenge because as he pointed out, the bigger you get, the harder it becomes to keep the growth trajectory going.
But the company has continued to thrive and is on track to pass $3 billion in revenue some time in the next couple of quarters. Red Hat is best known for creating a version of Linux designed specifically for the enterprise, but it has begun adapting to the changing world out there with cloud and containers — and as its RHEL (Red Hat Enterprise Linux) customers start to change the way they work (ever so slowly), they are continuing to use Red Hat for these new technologies. Whitehurst told me, that’s not a coincidence.
The cloud and containers are built on Linux, and if there is one thing Red Hat knows, it’s Linux. Whitehurst points out the legacy RHEL business is still growing at a healthy 14 percent, but it’s the newer cloud and container business that’s growing like gangbusters at a robust 40 percent, and he says that is really having a positive impact on revenue.
In its most recent earnings report last month, overall revenue was up 21% to $723 million for the quarter for a $2.8 billion run rate. Investors certainly seem to like what they are seeing. The share price has gone on straight upward trajectory from a low of $68.71 in December 2016 to $121 per share today, as I wrote this article. That’s a nice return any way you slice it.
Whitehurst says the different parts of the business are really feeding one another. The company made an early bet early on Kubernetes, the open source container orchestration tool originally developed at Google. That bet has paid off handsomely as companies are moving towards containerized application delivery using Kubernetes. In the same way Red Hat packaged Linux in a way that made sense for enterprise IT, it’s doing the same thing with Kubernetes with its OpenShift products. In fact, Whitehurst jokes OpenShift would be more widely recognized if they had just put Kubernetes in the name.
While he attributes some of the company’s success in this area in being in the right place at the right time with the right technology, he reckons it’s more than that. “We have some skill in identifying architecture that is best for the enterprise,” he said. It doesn’t hurt that they also got involved with contributing back to the community early on and today are second largest contributor to Kubernetes.
But he says the Linux connection, the fact that containers are built on Linux, is really what is the most likely factor driving the business, and that they can apply what they know in Linux to containers is a big deal.
But he points out that large organizations, who are his company’s bread and butter, aren’t all rushing to containerize their entire application inventory. These companies tend to move more slowly than that and Red Hat is trying to cover them regardless of where they are in that evolution: using virtual machines in the cloud or on prem or running containerized applications.
Whitehurst understands his company is selling free software, so they have to add value by easing the implementation and management of these tools for customers. “When you sell free software, you have to obsess about the value can bring because the IP is free,” he said. Given the numbers, it would appear customers see that value, and that is contributing to that steady march toward $5 billion.
Featured Image: Bloomberg/Getty Images
News Source = techcrunch.com
Google injects Hire with AI to speed up common tasks
Since Google Hire launched last year it has been trying to make it easier for hiring managers to manage the data and tasks associated with the hiring process, while maybe tweaking LinkedIn while they’re at it. Today the company announced some AI-infused enhancements that they say will help save time and energy spent on manual processes.
“By incorporating Google AI, Hire now reduces repetitive, time-consuming tasks, like scheduling interviews into one-click interactions. This means hiring teams can spend less time with logistics and more time connecting with people,” Google’s Berit Hoffmann, Hire product manager wrote in a blog post announcing the new features.
The first piece involves making it easier and faster to schedule interviews with candidates. This is a multi-step activity that involves scheduling appropriate interviewers, choosing a time and date that works for all parties involved in the interview and scheduling a room in which to conduct the interview. Organizing these kind of logistics tend to eat up a lot of time.
“To streamline this process, Hire now uses AI to automatically suggest interviewers and ideal time slots, reducing interview scheduling to a few clicks,” Hoffmann wrote.
Another common hiring chore is finding keywords in a resume. Hire’s AI now finds these words for a recruiter automatically by analysing terms in a job description or search query and highlighting relevant words including synonyms and acronyms in a resume to save time spent manually searching for them.
Finally, another standard part of the hiring process is making phone calls, lots of phone calls. To make this easier, the latest version of Google Hire has a new click-to-call function. Simply click the phone number and it dials automatically and registers the call in call a log for easy recall or auditing.
While Microsoft has LinkedIn and Office 365, Google has G Suite and Google Hire. The strategy behind Hire is to allow hiring personnel to work in the G Suite tools they are immersed in every day and incorporate Hire functionality within those tools.
It’s not unlike CRM tools that integrate with Outlook or GMail because that’s where sales people spend a good deal of their time anyway. The idea is to reduce the time spent switching between tools and make the process a more integrated experience.
While none of these features individually will necessarily wow you, they are making use of Google AI to simplify common tasks to reduce some of the tedium associated with every-day hiring tasks.
News Source = techcrunch.com
Amazon starts shipping its $249 DeepLens AI camera for developers
Back at its re:Invent conference in November, AWS announced its $249 DeepLens, a camera that’s specifically geared toward developers who want to build and prototype vision-centric machine learning models. The company started taking pre-orders for DeepLens a few months ago, but now the camera is actually shipping to developers.
Ahead of today’s launch, I had a chance to attend a workshop in Seattle with DeepLens senior product manager Jyothi Nookula and Amazon’s VP for AI Swami Sivasubramanian to get some hands-on time with the hardware and the software services that make it tick.
DeepLens is essentially a small Ubuntu- and Intel Atom-based computer with a built-in camera that’s powerful enough to easily run and evaluate visual machine learning models. In total, DeepLens offers about 106 GFLOPS of performance.
The hardware has all of the usual I/O ports (think Micro HDMI, USB 2.0, Audio out, etc.) to let you create prototype applications, no matter whether those are simple toy apps that send you an alert when the camera detects a bear in your backyard or an industrial application that keeps an eye on a conveyor belt in your factory. The 4 megapixel camera isn’t going to win any prizes, but it’s perfectly adequate for most use cases. Unsurprisingly, DeepLens is deeply integrated with the rest of AWS’s services. Those include the AWS IoT service Greengrass, which you use to deploy models to DeepLens, for example, but also SageMaker, Amazon’s newest tool for building machine learning models.
These integrations are also what makes getting started with the camera pretty easy. Indeed, if all you want to do is run one of the pre-built samples that AWS provides, it shouldn’t take you more than 10 minutes to set up your DeepLens and deploy one of these models to the camera. Those project templates include an object detection model that can distinguish between 20 objects (though it had some issues with toy dogs, as you can see in the image above), a style transfer example to render the camera image in the style of van Gogh, a face detection model and a model that can distinguish between cats and dogs and one that can recognize about 30 different actions (like playing guitar, for example). The DeepLens team is also adding a model for tracking head poses. Oh, and there’s also a hot dog detection model.
But that’s obviously just the beginning. As the DeepLens team stressed during our workshop, even developers who have never worked with machine learning can take the existing templates and easily extend them. In part, that’s due to the fact that a DeepLens project consists of two parts: the model and a Lambda function that runs instances of the model and lets you perform actions based on the model’s output. And with SageMaker, AWS now offers a tool that also makes it easy to build models without having to manage the underlying infrastructure.
You could do a lot of the development on the DeepLens hardware itself, given that it is essentially a small computer, though you’re probably better off using a more powerful machine and then deploying to DeepLens using the AWS Console. If you really wanted to, you could use DeepLens as a low-powered desktop machine as it comes with Ubuntu 16.04 pre-installed.
For developers who know their way around machine learning frameworks, DeepLens makes it easy to import models from virtually all the popular tools, including Caffe, TensorFlow, MXNet and others. It’s worth noting that the AWS team also built a model optimizer for MXNet models that allows them to run more efficiently on the DeepLens device.
So why did AWS build DeepLens? “The whole rationale behind DeepLens came from a simple question that we asked ourselves: How do we put machine learning in the hands of every developer,” Sivasubramanian said. “To that end, we brainstormed a number of ideas and the most promising idea was actually that developers love to build solutions as hands-on fashion on devices.” And why did AWS decide to build its own hardware instead of simply working with a partner? “We had a specific customer experience in mind and wanted to make sure that the end-to-end experience is really easy,” he said. “So instead of telling somebody to go download this toolkit and then go buy this toolkit from Amazon and then wire all of these together. […] So you have to do like 20 different things, which typically takes two or three days and then you have to put the entire infrastructure together. It takes too long for somebody who’s excited about learning deep learning and building something fun.”
So if you want to get started with deep learning and build some hands-on projects, DeepLens is now available on Amazon. At $249, it’s not cheap, but if you are already using AWS — and maybe even use Lambda already — it’s probably the easiest way to get started with building these kind of machine learning-powered applications.
News Source = techcrunch.com
Sumo Logic brings data analysis to containers
Sumo Logic has long held the goal to help customers understand their data wherever it lives. As we move into the era of containers, that goal becomes more challenging because containers by their nature are ephemeral. The company announced a product enhancement today designed to instrument containerized applications in spite of that.
Sumo’s CEO Ramin Sayer says containers have begun to take hold over the last 12-18 months with Docker and Kubernetes emerging as tools of choice. Given their popularity, Sumo wants to be able to work with them. “[Docker and Kubernetes] are by far the most standard things that have developed in any new shop, or any existing shop that wants to build a brand new modern app or wants to lift and shift an app from on prem [to the cloud], or have the ability to migrate workloads from Vendor A platform to Vendor B,” he said.
He’s not wrong of course. Containers and Kubernetes have been taking off in a big way over the last 18 months and developers and operations alike have struggled to instrument these apps to understand how they behave.
“But as that standardization of adoption of that technology has come about, it makes it easier for us to understand how to instrument, collect, analyze, and more importantly, start to provide industry benchmarks,” Sayer explained.
They do this by avoiding the use of agents. Regardless of how you run your application, whether in a VM or a container, Sumo is able to capture the data and give you feedback you might otherwise have trouble retrieving.
The company has built in native support for Kubernetes and Amazon Elastic Container Service for Kubernetes (Amazon EKS). It also supports the open source tool Prometheus favored by Kubernetes users to extract metrics and metadata. The goal of the Sumo tool is to help customers fix issues faster and reduce downtime.
As they work with this technology, they can begin to understand norms and pass that information onto customers. “We can guide them and give them best practices and tips, not just on what they’ve done, but how they compare to other users on Sumo,” he said.
News Source = techcrunch.com
Follow on Twitter
Delhi1 month ago
Chefow – Delhi based food tech startup to bring home cooked food at your doorstep.
Delhi1 month ago
A Start of New Technical Ecosystem by International Blockchain Technology Council
Delhi1 month ago
Army of Poets – Education For All
Delhi1 year ago
Vistaprint to shutdown Indian Operations? 75% Employees Fired, Senior Leaders
Amazing8 months ago
वीडियो : सलमान खान की शादी हो या ना हो लेकिन राहुल गांधी ने अपनी शादी को लेकर खोला राज
Delhi1 month ago
Dheeraj Sharma, National President Of NSC to partner with DIYguru for NSC Students
India1 year ago
How BajaTutor is helping Car Racing enthusiasts to build their ATV’s?
Education1 year ago
Haridwar Boy making his way to promote Online Learning in India for GATE / IES Exams