The growing popularity of visual applications for displaying increasingly rich data sets is clearly a key driver behind the rising momentum of Android among embedded developers. Industrial Automation (28%), Infotainment (20%), and Digital Signage (12%) were the top three applications in the survey, closely followed by HMI (11%) and Medical (11%).
The ability to add a touch interface (26%) was listed by respondents as the main benefit of adopting Android, with reduced time to market (25%) and customizability (19%) coming in second and third respectively. Native multimedia support (14%) in the O/S and the robust Android app ecosystem (12%) were also seen as important.
As a mobile O/S, Android does offer some critical challenges for developers to overcome when implementing it for embedded applications, including its lack of I/O support for peripherals (23%), the need to maintain different versions of it, and ensuring security (17%). Building up internal Android development expertise (21%) and finding the right development tools (19%) are other key issues embedded developers face in adopting the O/S.
We’d like to thank everyone who responded to the survey for the invaluable feedback they provided. We have already begun analyzing the results in order to determine how we can improve the platforms and services we offer for Embedded Android, and will be updating you on our progress.
Download the full results here.
Home » intelligent systems
Category Archives: intelligent systems
DataStax: a fully distributed and highly secure transactional database platform that is “always on”
When an open-source database written in Java that runs primarily in production on Linux becomes THE solution for the cloud platform from Microsoft (i.e. Azure) in the fully distributed, highly secure and “always on” transactional database space then we should take a special note of that. This is the case of DataStax:
July 15, 2015: Building the intelligent cloud Scott Guthrie’s keynote on the Microsoft Worldwide Partner Conference 2015, the DataStax related segment in 7 minutes only
SCOTT GUTHRIE, EVP of Microsoft Cloud and Enterprise: What I’d like to do is invite three different partners now on stage, one an ISV, one an SI, and one a managed service provider to talk about how they’re taking advantage of our cloud offerings to accelerate their businesses and make their customers even more successful.
First, and I think, you know, being able to take advantage of all of these different capabilities that we now offer.
Now, the first partner I want to bring on stage is DataStax. DataStax delivers an enterprise-grade NoSQL offering based on Apache Cassandra. And they enable customers to build solutions that can scale across literally thousands of servers, which is perfect for a hyper-scale cloud environment.
And one of the customers that they’re working with is First American, who are deploying a solution on Microsoft Azure to provide richer insurance and settlement services to their customers.
What I’d like to do is invite Billy Bosworth, the CEO of DataStax, on stage to join me to talk about the partnership that we’ve had and how some of the great solutions that we’re building together. Here’s Billy. (Applause.)
Well, thanks for joining me, Billy. And it’s great to have you here.
BILLY BOSWORTH, CEO of DataStax: Thank you. It’s a real privilege to be here today.
SCOTT GUTHRIE: So tell us a little bit about DataStax and the technology you guys build.
BILLY BOSWORTH: Sure. At DataStax, we deliver Apache Cassandra in a database platform that is really purpose-built for the new performance and availability demands that are being generated by today’s Web, mobile and IOT applications.
With DataStax Enterprise, we give our customers a fully distributed and highly secure transactional database platform.
Now, that probably sounds like a lot of other database vendors out there as well. But, Scott, we have something that’s really different and really important to us and our customers, and that’s the notion of being always on. And when you talk about “always on” and transactional databases, things can get pretty complicated pretty fast, as you well know.
The reason for that is in an always-on world, the datacenter itself becomes a single point of failure. And that means you have to build an architecture that is going to be comprehensive and include multiple datacenters. That’s tough enough with almost any other piece of the software stack. But for transactional databases, that is really problematic.
Fortunately, we have a masterless architecture in Apache Cassandra that allows us to have DataStax enterprise scale in a single datacenter or across multiple datacenters, and yet at the same time remain operationally simple. So that’s really the core of what we do.
SCOTT GUTHRIE: Is the always-on angle the key differentiator in terms of the customer fit with Azure?
BILLY BOSWORTH: So if you think about deployment to multiple datacenters, especially and including Azure, it creates an immediate benefit. Going back to your hybrid clouds comment, we see a lot of our customers that begin their journey on premises. So they take their local datacenter, they install DataStax Enterprise, it’s an active database up and running. And then they extend that database into Azure.
Now, when I say that, I don’t mean they do so for disaster recovery or failover, it is active everywhere. So it is taking full read-write requests on premises and in Azure at the same time.
So if you lose connectivity to your physical datacenter, then the Azure active nodes simply take over. And that’s great, and that solves the always-on problem.
But that’s not the only thing that Azure helps to solve. Our applications, because of their nature, tend to drive incredibly high throughput. So for us, hundreds of millions or even tens and hundreds of billions of transactions a day is actually quite common.
You guys are pretty good, Scott, but I don’t think you’ve changed the laws of physics yet. And so the way that you get that kind of throughput with unbelievable performance demands, because our customers demand millisecond and microsecond response times, is you push the data closer to the end points. You geographically distribute it.
Now, what our customers are realizing is they can try and build 19 datacenters across the world, which I’m sure was really cheap and easy to do, or they can just look at what you’ve already done and turn to a partnership like ours to say, “Help us understand how we do this with Azure.”
So not only do you get the always-on benefit, which is critical, but there’s also a very important performance element to this type of architecture as well.
SCOTT GUTHRIE: Can you tell us a little bit about the work you did with First American on Azure?
BILLY BOSWORTH: Yeah. First American is a leading name in the title insurance and settlement services businesses. In fact, they manage more titles on more properties than anybody in the world.
Every title comes with an associated set of metadata. And that metadata becomes very important in the new way that they want to do business because each element of that needs to be transacted, searched, and done in real-time analysis to provide better information back to the customer in real time.
And so for that on the database side, because of the type of data and because of the scale, they needed something like DataStax Enterprise, which we’ve delivered. But they didn’t want to fight all those battles of the architecture that we discussed on their own, and that’s where they turned to our partnership to incorporate Microsoft Azure as the infrastructure with DataStax Enterprise running on top.
And this is one of many engagements that you know we have going on in the field that are really, really exciting and indicative of the way customers are thinking about transforming their business.
SCOTT GUTHRIE: So what’s it like working with Microsoft as a partner?
BILLY BOSWORTH: I tell you, it’s unbelievable. Or, maybe put differently, highly improbable that you and I are on stage together. I want you guys to think about this. Here’s the type of company we are. We’re an open-source database written in Java that runs primarily in production on Linux.
Now, Scott, Microsoft has a couple of pretty good databases, of which I’m very familiar from my past, and open source and Java and Linux haven’t always been synonymous with Microsoft, right?
So I would say the odds of us being on stage were almost none. But over the past year or two, the way that you guys have opened up your aperture to include technologies like ours — and I don’t just say “include.” His team has embraced us in a way that is truly incredible. For a company the size of Microsoft to make us feel the way we do is just remarkable given the fact that none of our technologies have been something that Microsoft has traditionally said is part of their family.
So I want to thank you and your team for all the work you’ve done. It’s been a great experience, but we are architecting systems that are going to drive businesses for the coming decades. And that is super exciting to have a partner like you engaged with us.
SCOTT GUTHRIE: Fantastic. Well, thank you so much for joining us on stage.
BILLY BOSWORTH: Thanks, Scott. (Applause.)
The typical data framework capabilities of DataStax in all respects is best understood via the the following webinar which presents Apache Spark as well as the part of the complete data platform solution:
– Apache Cassandra is the leading distributed database in use at thousands of sites with the world’s most demanding scalability and availability requirements.
– Apache Spark is a distributed data analytics computing framework that has gained a lot of traction in processing large amounts of data in an efficient and user-friendly manner.
– The joining of both provides a powerful combination of real-time data collection with analytics.
After a brief overview of Cassandra and Spark, (Cassandra till 16:39, Spark till 19:25) this class will dive into various aspects of the integration (from 19:26).
August 19, 2015: Big Data Analytics with Cassandra and Spark by Brian Hess, Senior Product Manager of Analytics, DataStax
September 23, 2015: DataStax Announces Strategic Collaboration with Microsoft, company press release
- DataStax delivers a leading fully-distributed database for public and private cloud deployments
- DataStax Enterprise on Microsoft Azure enables developers to develop, deploy and monitor enterprise-ready IoT, Web and mobile applications spanning public and private clouds
- Scott Guthrie, EVP Cloud and Enterprise, Microsoft, to co-deliver Cassandra Summit 2015 keynote
SANTA CLARA, CA – September 23, 2015 – (Cassandra Summit 2015) DataStax, the company that delivers Apache Cassandra™ to the enterprise, today announced a strategic collaboration with Microsoft to deliver Internet of Things (IoT), Web and mobile applications in public, private or hybrid cloud environments. With DataStax Enterprise (DSE), a leading fully-distributed database platform, available on Azure, Microsoft’s cloud computing platform, enterprises can quickly build high-performance applications that can massively scale and remain operationally simple across public and private clouds, with ease and at lightning speed.
Click to Tweet: #DataStax Announces Strategic Collaboration with @Microsoft at #CassandraSummit bit.ly/1V8KY4D
PERSPECTIVES ON THE NEWS
“At Microsoft we’re focused on enabling customers to run their businesses more productively and successfully,” said Scott Guthrie, Executive Vice President, Cloud and Enterprise, Microsoft. “As more organizations build their critical business applications in the cloud, DataStax has proved to be a natural Azure partner through their ability to enable enterprises to build solutions that can scale across thousands of servers which is necessary in today’s hyper-scale cloud environment.”
“We are witnessing an increased adoption of DataStax Enterprise deployments in hybrid cloud environments, so closely aligning with Microsoft benefits any organization looking to quickly and easily build high-performance IoT, Web and mobile apps,” said Billy Bosworth, CEO, DataStax. “Working with a world-class organization like Microsoft has been an incredible experience and we look forward to continuing to work together to meet the needs of enterprises looking to successfully transition their business to the cloud.”
“As a leader in providing information and insight in critical areas that shape today’s business landscape, we knew it was critical to transform our back-end business processes to address scale and flexibility” said Graham Lammers, Director, IHS. “With DataStax Enterprise on Azure we are now able to create a next generation big data application to support the decision-making process of our customers across the globe.”
BUILD SIMPLE, SCALABLE AND ALWAY-ON APPS IN RECORD SPEED
To address the ever-increasing demands of modern businesses transitioning from on-premise to hybrid cloud environments, the DataStax Enterprise on Azure on-demand cloud database solution provides enterprises with both development and production ready Bring Your Own License (BYOL) DSE clusters that can be launched in minutes on theMicrosoft Azure Marketplace using Azure Resource Management (ARM) Templates. This enables the building of high-performance IoT, Web and mobile applications that can predictably scale across global Azure data centers with ease and at remarkable speed. Additional benefits include:
- Hybrid Deployment: Easily move DSE workloads between data centers, service providers and Azure, and build hybrid applications that leverage resources across all three.
- Simplicity: Easily manage, develop, deploy and monitor database clusters by eliminating data management complexities.
- Scalability: Quickly replicate online applications globally across multiple data centers into the cloud/hybrid cloud environment.
- Continuous Availability: DSE’s peer-to-peer architecture offers no single point of failure. DSE also provides maximum flexibility to distribute data where it’s needed most by replicating data across multiple data centers, the cloud and mixed cloud/on-premise environments.
MICROSOFT ENTERPRISE CLOUD ALLIANCE & FAST START PROGRAM
DataStax also announced it has joined Microsoft’s Enterprise Cloud Alliance, a collaboration that reinforces DataStax’scommitment to provide the best set of on-premise, hosted and public cloud database solutions in the industry. The goal of Microsoft’s Enterprise Cloud Alliance partner program is to create, nurture and grow a strong partner ecosystem across a broad set of Enterprise Cloud Products delivering the best on-premise, hosted and Public Cloud solutions in the industry. Through this alliance, DataStax and Microsoft are working together to create enhanced enterprise-grade offerings for the Azure Marketplace that reduce the complexities of deployment and provisioning through automated ARM scripting capabilities.
Additionally, as a member of Microsoft Azure’s Fast Start program, created to help users quickly deploy new cloud workloads, DataStax users receive immediate access to the DataStax Enterprise Sandbox on Azure for a hands-on experience testing out DSE on Azure capabilities. DataStax Enterprise Sandbox on Azure can be found here.
Cassandra Summit 2015, the world’s largest gathering of Cassandra users, is taking place this week and Microsoft Cloud and Enterprise Executive Vice President Scott Guthrie, DataStax CEO Billy Bosworth, and Apache Cassandra Project Chair and DataStax Co-founder and CTO Jonathan Ellis, will deliver the conference keynote at 10 a.m. PT on Wednesday, September 23. The keynote can be viewed at DataStax.com.
ABOUT DATASTAX
DataStax delivers Apache Cassandra™ in a database platform purpose-built for the performance and availability demands for IoT, Web and mobile applications. This gives enterprises a secure, always-on database technology that remains operationally simple when scaling in a single datacenter or across multiple datacenters and clouds.
With more than 500 customers in over 50 countries, DataStax is the database technology of choice for the world’s most innovative companies, such as Netflix, Safeway, ING, Adobe, Intuit and eBay. Based in Santa Clara, Calif., DataStax is backed by industry-leading investors including Comcast Ventures, Crosslink Capital, Lightspeed Venture Partners, Kleiner Perkins Caufield & Byers, Meritech Capital, Premji Invest and Scale Venture Partners. For more information, visit DataStax.com or follow us @DataStax.
September 30, 2014: Why Datastax’s increasing presence threatens Oracle’s database by Anne Shields at Market Realist
Must know: An in-depth review of Oracle’s 1Q15 earnings (Part 9 of 12)
Datastax databases are built on open-source technologies
Datastax is a California-based database management company. It offers an enterprise-grade NoSQL database that seamlessly and securely integrates real-time data with Apache Cassandra. Databases built on Apache Cassandra offer more flexibility than traditional databases. Even in case of calamities and uncertainties, like floods and earthquakes, data is available due to its replication at other data centers. NoSQL and Cassandra are open-source software.
Cassandra database was developed by Facebook (FB) to handle its enormous volumes of data. The technology behind Cassandra was developed by Amazon (AMZN) and Google (GOOGL). Oracle’s MySQL (ORCL), Microsoft’s SQL Server (MSFT), and IBM’s DB2 (IBM) are the traditional databases present in the market .
The above chart shows how NoSQL databases, NewSQL databases, and Data grid/cache products fit into the wider data management landscape.
Huge amounts of funds raised in the open-source technology database space
Datastax raised $106 million in September 2014 to expand its database operations. MongoDB Inc. and Couchbase Inc.—both open-source NoSQL database developers—raised $231 million and $115 million, respectively, in 2014. According to Market Research Media, a consultancy firm, spending on NoSQL technology in 2013 was less than $1 billion. It’s expected to reach $3.4 billion by 2020. This explains why this segment is attracting such huge investments.
Oracle’s dominance in the database market is uncertain
Oracle claims it’s a market leader in the relational database market, with a revenue share of 48.3%. In 2013, it launched Oracle Database 12C. According to Oracle, “Oracle Database 12c introduces a new multitenant architecture that simplifies the process of consolidating databases onto the cloud; enabling customers to manage many databases as one — without changing their applications.” To know in detail about Database 12c, please click here .
In July 2013, DataStax announced that dozens of companies have migrated from Oracle databases to DataStax databases. Customers cited scalability, disaster avoidance, and cost savings as the reasons for shifting databases. Datastax databases’ rising popularity jeopardizes Oracle’s dominant position in the database market.
Browse this series on Market Realist:
September 24, 2014: Building a better experience for Azure and DataStax customers by Matt Rollender, VP Cloud Strategy, DataStax, Inc. on Microsoft Azure blog
Cassandra Summit is in high gear this week in Santa Clara, CA, representing the largest NoSQL event of its kind! This is the largest Cassandra Summit to date. With more than 7,000 attendees (both onsite and virtual), this is the first time the Summit is a three-day event with over 135 speaking sessions. This is also the first time DataStax will debut a formalized Apache Cassandra™ training and certification program in conjunction with O’Reilly Media. All incredibly exciting milestones!
We are excited to share another milestone. Yesterday, we announced our formal strategic collaboration with Microsoft. Dedicated DataStax and Microsoft teams have been collaborating closely behind the scenes for more than a year on product integration, QA testing, platform optimization, automated provisioning, and characterization of DataStax Enterprise (DSE) on Azure, and more to ensure product validation and a great customer experience for users of DataStax Enterprise on the Azure cloud. There is strong coordination across the two organizations – very close executive, field, and technical alignment – all critical components for a strong partnership.
This partnership is driven and shaped by our joint customers. Our customers oftentimes begin their journey with on-premise deployments of our database technology and then have a requirement to move to the cloud – Microsoft is a fantastic partner to help provide the flexibility of a true hybrid environment along with the ability to migrate to and scale applications in the cloud. Additionally, Microsoft has significant breadth regarding their data centers – customers can deploy in numerous Azure data centers around the globe, in order to be ‘closer’ to their end users. This is highly complementary to DataStax Enterprise software as we are a peer-to-peer distributed database and our customers need to be close to their end users with their always-on, always available enterprise applications.
To highlight a couple of joint customers and use cases we have First American Title and IHS, Inc. First American is a leading provider of title insurance and settlement services with revenue over $5B. They ingest and store the largest number (billions) of real estate property records in the industry. Accessing, searching and analyzing large data-sets to get relevant details quickly is the new way they want to do business – to provide better information back to their customers in real-time and allow end users to easily search through the property records on-line. They chose DSE and Azure because of the large data requirements and because of the need to continue to scale the application.
A second great customer and use case is IHS, Inc., a $2B revenue-company that provides information and analysis to support the decision-making process of businesses and governments. This is a transformational project for IHS as they are building out an ‘internet age’ parts catalog – it’s a next generation big data application, using NoSQL, non-relational technology and they want to deploy in the cloud to bring the application to market faster.
As you can see, we are enabling enterprises to engage their customer like never before with their always on, highly available and distributed applications. Stay tuned for more as we move forward together in the coming months!
For Additional information go to http://www.datastax.com/marketplace-microsoft-azure to try out Datastax Enterprise Sandbox on Azure.
See also DataStax Enterprise Cluster Production on Microsoft Azure Marketplace
September 23, 2015: Making Cassandra Do Azure, But Not Windows by Timothy Prickett Morgan Co-Editor, Co-Founder, The Next Platform
When Microsoft says that it is embracing Linux as a peer to Windows, it is not kidding. The company has created its own Linux distribution for switches used to build the Azure cloud, and it has embraced Spark in-memory processing and Cassandra as its data store for its first major open source big data project – in this case to help improve the quality of its Office365 user experience. And now, Microsoft is embracing Cassandra, the NoSQL data store originally created by Facebook when it could no longer scale the MySQL relational database to suit its needs, on the Azure public cloud.
Billy Bosworth, CEO at DataStax, the entity that took over steering development of and providing commercial support for Cassandra, tells The Next Platform that the deal with Microsoft has a number of facets, all of which should help boost the adoption of the enterprise-grade version of Cassandra. But the key one is that the Global 2000 customers that DataStax wants to sell support and services to are already quite familiar with both Windows Server in their datacenters and they are looking to burst out to the Azure cloud on a global scale.
“We are seeing a rapidly increasing number of our customers who need hybrid cloud, keeping pieces of our DataStax Enterprise on premise in their own datacenters and they also want to take pieces of that same live transactional data – not replication, but live data – and in the Azure cloud as well,” says Bosworth. “They have some unique capabilities, and one of the major requirements of customers is that even if they use cloud infrastructure, it still has to be distributed by the cloud provider. They can’t just run Cassandra in one availability zone in one region. They have to span data across the globe, and Microsoft has done a tremendous job of investing in its datacenters.”
With the Microsoft agreement, DataStax is now running its wares on the three big clouds, with Amazon Web Services and Google Compute Engine already certified able to run the production-grade Cassandra. And interestingly enough, Microsoft is supporting the DataStax implementation of Cassandra on top of Linux, not Windows. Bosworth says that while Cassandra can be run on Windows servers, DataStax does not recommend putting DataStax Enterprise (DSE), the commercial release, on Windows. (It does have a few customers who do, nonetheless, and it supports them.) Bosworth adds that DataStax and the Cassandra community have been “working diligently” for the past year to get a Windows port of DSE completed and that there has been “zero pressure” for the Microsoft Azure team to run DSE on anything other than Linux.
It is important to make the distinction between running Cassandra and other elements of DSE on Windows and having optimized drivers for Cassandra for the .NET programming environment for Windows.
“All we are really talking about is the ability to run the back-end Cassandra on Linux or Windows, and to the developer, it is irrelevant on what that back end is running,” explains Bosworth. This takes away some of that friction, and what we find is that on the back end, we just don’t find religious conviction about whether it should run on Windows or Linux, and this is different from five years ago. We sell mostly to enterprises, and we have not had one customer raise their hand and say they can’t use DSE because it does not run on Windows.”
What is more important is the ability to seamless put Cassandra on public clouds and spread transactional data around for performance and resiliency reasons – the same reasons that Facebook created Cassandra for in the first place.
What Is In The Stack, Who Uses It, And How
The DataStax Enterprise distribution does not just include the Apache Cassandra data store, but has an integrated search engine that is API compatible with the open source Solr search engine and in-memory extensions that can speed up data accesses by anywhere from 30X to 100X compared to server clusters using flash SSDs or disk drives. The Cassandra data store can be used to underpin Hadoop, allowing it to be queried by MapReduce, Hive, Pig, and Mahout, and it can also underpin Spark and Spark Streaming as their data stores if customers decide to not go with the Hadoop Distributed File System that is commonly packaged with a Hadoop distribution.
It is hard to say for sure how many organizations are running Cassandra today, but Bosworth reckons that it is on the order of tens of thousands worldwide, based on a number of factors. DataStax does not do any tracking of its DataStax Community edition because it wants a “frictionless download” like many open source projects have. (Developers don’t want software companies to see what tools they are playing with, even though they might love open source code.) DataStax provides free training for Cassandra, however, where it does keep track, and developers are consuming over 10,000 units of this training per month, so that probably indicates that the Cassandra installed base (including tests, prototypes, and production) is in the five figures.
DataStax itself has over 500 paying customers – now including Microsoft after its partner tried to build its own Spark-Cassandra cluster using open source code and decided that the supported versions were better thanks to the extra goodies that DataStax puts into its distro. DataStax has 30 of the Fortune 100 using its distribution of Cassandra in one form or another, and it is always for transactional, rather than batch analytic, jobs and in most cases also for distributed data stores that make use of the “eventual consistency” features of Cassandra to replicate data across multiple clusters. The company has another 600 firms participating in its startup program, which gives young companies freebie support on the DSE distro until they hit a certain size and can afford to start kicking some cash into the kitty.
The largest installation of Cassandra is running at Apple, which as we previously reported has over 75,000 nodes, with clusters ranging in size from hundreds to over 1,000 nodes and with a total capacity in the petabytes range. Netflix, which used to employ the open source Cassandra, switched to DSE last May and had over 80 clusters with more than 2,500 nodes supporting various aspects of its video distribution business. In both cases, Cassandra is very likely housing user session state data as well as feeding product or play lists and recommendations or doing faceted search for their online customers.
We are always intrigued to learn how customers are actually deploying tools such as Cassandra in production and how they scale it. Bosworth says that it is not uncommon to run a prototype project on as few as ten nodes, and when the project goes into production, to see it grow to dozens to hundreds of nodes. The midrange DSE clusters range from maybe 500 to 1,000 nodes and there are some that get well over 1,000 nodes for large-scale workloads like those running at Apple.
In general, Cassandra does not, like Hadoop, run on disk-heavy nodes. Remember, the system was designed to support hot transactional data, not to become a lake with a mix of warm and cold data that would be sifted in batch mode as is still done with MapReduce running atop Hadoop.
The typical node configuration has changed as Cassandra has evolved and improved, says Robin Schumacher, vice president of products at DataStax. But before getting into feeds and speeds, Schumacher offered this advice. “There are two golden rules for Cassandra. First, get your data model right, and second, get your storage system right. If you get those two things right, you can do a lot wrong with your configuration or your hardware and Cassandra will still treat you right. Whenever we have to dive in and help someone out, it is because they have just moved over a relational data model or they have hooked their servers up to a NAS or a SAN or something like that, which is absolutely not recommended.”
Only four years ago, because of the limitations in Cassandra (which like Hadoop and many other analytics tools is coded in Java), the rule of thumb was to put no more than 512 GB of disk capacity onto a single node. (It is hard to imagine such small disk capacities these days, with 8 TB and 10 TB disks.) The typical Cassandra node has two processors, with somewhere between 12 and 24 cores, and has between 64 GB and 128 GB of main memory. Customers who want the best performance tend to go with flash SSDs, although you can do all-disk setups, too.
Fast forward to today, and Cassandra can make use of a server node with maybe 5 TB of capacity for a mix of reads and writes, and if you have a write intensive application, then you can push that up to 20 TB. (DataStax has done this in its labs, says Schumacher, without any performance degradation.) Pushing the capacity up is important because it helps reduce server node count for a given amount of storage, which cuts hardware and software licensing and support costs. Incidentally, only a quarter of DSE customers surveyed said they were using spinning disks, but disk drives are fine for certain kinds of log data. SSDs are used for most transactional data, but the bits that are most latency sensitive should use DSE to store data on PCI-Express flash cards, which have lower latency.
Schumacher says that in most cases, the commercial-grade DSE Cassandra is used for a Web or mobile application, and a DSE cluster is not set up for hosting multiple applications, but rather companies have a different cluster for each use case. (As you can see is the case with Apple and Netflix.) Most of the DSE shops to make use of the eventual consistency replication features of Cassandra to span multiple datacenters with their data stores, and span anywhere from eight to twelve datacenters with their transactional data.
Here’s where it gets interesting, and why Microsoft is relevant to DataStax. Only about 30 percent of the DSE installations are running on premises. The remaining 70 percent are running on public clouds. About half of DSE customers are running on Amazon Web Services, with the remaining 20 percent split more or less evenly between Google Compute Engine and Microsoft Azure. If DataStax wants to grow its business, the easiest way to do that is to grow along with AWS, Compute Engine, and Azure.
So Microsoft and DataStax are sharing their roadmaps and coordinating development of their respective wares, and will be doing product validation, benchmarking, and optimization. The two will be working on demand generation and marketing together, too, and aligning their compensation to sell DSE on top of Azure and, eventually, on top of Windows Server for those who want to run it on premises.
In addition to announcing the Microsoft partnership at the Cassandra Summit this week, DataStax is also releasing its DSE 4.8 stack, which includes certification for Cassandra to be used as the back end for the new Spark 1.4 in-memory analytics tool. DSE Search has a performance boosts for live indexing, and running DSE instances inside of Docker containers has been improved. The stack also includes Titan 1.0, the graph database overlay for Cassandra, HBase, and BerkeleyDB that DataStax got through its acquisition of Aurelius back in February. DataStax is also previewing Cassandra 3.0, which will include support for JSON documents, role-based access control, and a lot of little tweaks that will make the storage more efficient, DataStax says. It is expected to ship later this year.
Satya Nadella on “Digital Work and Life Experiences” supported by “Cloud OS” and “Device OS and Hardware” platforms–all from Microsoft
Update: Gates Says He’s Very Happy With Microsoft’s Nadella [Bloomberg TV, Oct 2, 2014] + Bill Gates is trying to make Microsoft Office ‘dramatically better’ [The Verge, Oct 3, 2014]
This is the essence of Microsoft Fiscal Year 2014 Fourth Quarter Earnings Conference Call(see also the Press Release and Download Files) for me, as the new, extremely encouraging, overall setup of Microsoft in strategic terms (the below table is mine based on what Satya Nadella told on the conference call):
These are extremely encouraging strategic advancements vis–à–vis previously publicized ones here in the following, Microsoft related posts of mine:
- Microsoft Surface Pro 3 is the ultimate tablet product from Microsoft. What the market response will be? [this same blog, May 21, 2014]
- What Microsoft will do with the Nokia Devices and Services now taken over, but currently producing a yearly loss rate of as much as $1.5 billion? [this same blog, April 29, 2014]
- Microsoft BUILD 2014 Day 2: “rebranding” to Microsoft Azure and moving toward a comprehensive set of fully-integrated backend services [this same blog, April 27, 2014]
- Microsoft is transitioning to a world with more usage and more software driven value add (rather than the old device driven world) in mobility and the cloud, the latter also helping to grow the server business well above its peers [this same blog, April 25, 2014]
- Intel’s desperate attempt to establish a sizeable foothold on the tablet market until its 14nm manufacturing leadership could provide a profitable position for the company in 2016 [this same blog, April 27, 2014]
- Intel CTE initiative: Bay Trail-Entry V0 (Z3735E and Z3735D) SoCs are shipping next week in $129 Onda (昂达) V819i Android tablets—Bay Trail-Entry V2.1 (Z3735G and Z3735F) SoCs might ship in $60+ Windows 8.1 tablets from Emdoor Digital (亿道) in the 3d quarter [this same blog, April 11, 2014]
- Enhanced cloud-based content delivery services to anyone, on any device – from Microsoft (Microsoft Azure Media Services) and its solution partners [this same blog, April 8, 2014]
- Microsoft BUILD 2014 Day 1: new and exciting stuff for MS developers [this same blog, April 5, 2014]
- IDF14 Shenzhen: Intel is levelling the Wintel playing field with Android-ARM by introducing new competitive Windows tablet price points from $99 – $129 [this same blog, April 4, 2014]
- Microsoft BUILD 2014 Day 1: consistency and superiority accross the whole Windows family extended now to TVs and IoT devices as well—$0 royalty licensing program for OEM and ODM partners in sub 9” phone and tablet space [this same blog, April 2, 2014]
- An upcoming new era: personalised, pro-active search and discovery experiences for Office 365 (Oslo) [this same blog, April 2, 2014]
- OneNote is available now on every platform (+free!!) and supported by cloud services API for application and device builders [this same blog, March 18, 2014]
- View from Redmond via Tim O’Brien, GM, Platform Strategy at Microsoft [this same blog, Feb 21, 2014]
- “Cloud first”: the origins and the current meaning [this same blog, Feb 18, 2014]
- “Mobile first”: the origins and the current meaning [this same blog, Feb 18, 2014]
- Microsoft’s half-baked cloud computing strategy (H1’FY14) [this same blog, Feb 17, 2014]
- The first “post-Ballmer” offering launched: with Power BI for Office 365 everyone can analyze, visualize and share data in the cloud [this same blog, Feb 10, 2014]
- John W. Thompson, Chairman of the Board of Microsoft: the least recognized person in the radical two-men shakeup of the uppermost leadership [this same blog, Feb 6, 2014]
- The extraordinary attempt by Nokia/Microsoft to crack the U.S. market in terms of volumes with Nokia Lumia 521 (with 4G/LTE) and Nokia Lumia 520 [this same blog, Jan 18, 2014]
- 2014 will be the last year of making sufficient changes for Microsoft’s smartphone and tablet strategies, and those changes should be radical if the company wants to succeed with its devices and services strategy [this same blog, Jan 17, 2014]
- Will, with disappearing old guard, Satya Nadella break up the Microsoft behemoth soon enough, if any? [this same blog, Feb 5, 2014]
- Microsoft products for the Cloud OS [this same blog, Dec 18, 2013]
- Satya Nadella’s (?the next Microsoft CEO?) next ten years’ vision of “digitizing everything”, Microsoft opportunities and challenges seen by him with that, and the case of Big Data [this same blog, Dec 13, 2013]
- Leading PC vendors of the past: Go enterprise or die! [this same blog, Nov 7, 2013]
- Microsoft could be acquired in years to come by Amazon? The joke of the day, or a certain possibility (among other ones)? [this same blog, Sept 16, 2013]
- The question mark over Wintel’s future will hang in the air for two more years [this same blog, Sept 15, 2013]
- The long awaited Windows 8.1 breakthrough opportunity with the new Intel “Bay Trail-T”, “Bay Trail-M” and “Bay Trail-D” SoCs? [this same blog, Sept 14, 2013]
- How the device play will unfold in the new Microsoft organization? [this same blog, July 14, 2013]
- Microsoft reorg for delivering/supporting high-value experiences/activities [this same blog, July 11, 2013]
- Microsoft partners empowered with ‘cloud first’, high-value and next-gen experiences for big data, enterprise social, and mobility on wide variety of Windows devices and Windows Server + Windows Azure + Visual Studio as the platform [this same blog, July 10, 2013]
- Windows Azure becoming an unbeatable offering on the cloud computing market [this same blog, June 28, 2013]
- Proper Oracle Java, Database and WebLogic support in Windows Azure including pay-per-use licensing via Microsoft + the same Oracle software supported on Microsoft Hyper-V as well [this same blog, June 25, 2013]
- “Cloud first” from Microsoft is ready to change enterprise computing in all of its facets [this same blog, June 4, 2013]
I see, however, particularly challenging the continuation of the Lumia story with the above strategy, as with the previous, combined Ballmer/Elop(Nokia) strategy the results were extremely weak:
Worthwhile to include here the videos Bloomberg was publishing simultaneously with Microsoft Fourth Quarter Earnings Conference Call:
Inside Microsoft’s Secret Surface Labs [Bloomberg News, July 22, 2014]
Will Microsoft Kinect Be a Medical Game-Changer? [Bloomberg News, July 22, 2014]
Why Microsoft Puts GPS In Meat For Alligators [Bloomberg News, July 22, 2014]
To this it is important to add: How Pier 1 is using the Microsoft Cloud to build a better relationship with their customers [Microsoft Server and Cloud YouTube channel, July 21, 2014]
as well as:
Microsoft Surface Pro 3 vs. MacBook Air 13″ 2014 [CNET YouTube channel, July 21, 2014]
Surface Pro 3 vs. MacBook Air (2014) [CTNtechnologynews YouTube channel, July 1, 2014]
In addition here are some explanatory quotes (for the new overall setup of Microsoft) worth to include here from the Q&A part of Microsoft’s (MSFT) CEO Satya Nadella on Q4 2014 Results – Earnings Call Transcript [Seeking Alpha, Jul. 22, 2014 10:59 PM ET]
…
Mark Moerdler – Sanford Bernstein
Thank you. And Amy one quick question, we saw a significant acceleration this quarter in cloud revenue, or I guess Amy or Satya. You saw acceleration in cloud revenue year-over-year what’s – is this Office for the iPad, is this Azure, what’s driving the acceleration and how long do you think we can keep this going?
Mark, I will take it and if Satya wants to add, obviously, he should do that. In general, I wouldn’t point to one product area. It was across Office 365, Azure and even CRM online. I think some of the important dynamics that you could point to particularly in Office 365; I really think over the course of the year, we saw an acceleration in moving the product down the market into increasing what we would call the mid-market and even small business at a pace. That’s a particular place I would tie back to some of the things Satya mentioned in the answer to your first question.
Improvements to analytics, improvements to understanding the use scenarios, improving the product in real-time, understanding trial ease of use, ease of sign-up all of these things actually can afford us the ability to go to different categories, go to different geos into different segments. And in addition, I think what you will see more as we initially moved many of our customers to Office 365, it came on one workload. And I think what we’ve increasingly seen is our ability to add more workloads and sell the entirety of the suite through that process. I also mentioned in Azure, our increased ability to sell some of these higher value services. So while, I can speak broadly but all of them, I think I would generally think about the strength of being both completion of our product suite ability to enter new segments and ability to sell new workloads.
The only thing I would add is it’s the combination of our SaaS like Dynamics in Office 365, a public cloud offering in Azure. But also our private and hybrid cloud infrastructure which also benefits, because they run on our servers, cloud runs on our servers. So it’s that combination which makes us both unique and reinforcing. And the best example is what we are doing with Azure active directory, the fact that somebody gets on-boarded to Office 365 means that tenant information is in Azure AD that fact that the tenant information is in Azure AD is what makes EMS or our Enterprise Mobility Suite more attractive to a customer manager iOS, Android or Windows devices. That network effect is really now helping us a lot across all of our cloud efforts.
…
Keith Weiss – Morgan Stanley
Excellent, thank you for the question and a very nice quarter. First, I think to talk a little bit about the growth strategy of Nokia, you guys look to cut expenses pretty aggressively there, but this is – particularly smartphones is a very competitive marketplace, can you tell us a little bit about sort of the strategy to how you actually start to gain share with Lumia on a going forward basis? And may be give us an idea of what levels of share or what levels of kind unit volumes are you going to need to hit to get to that breakeven in FY16?
Let me start and Amy you can even add. So overall, we are very focused on I would say thinking about mobility share across the entire Windows family. I already talked about in my remarks about how mobility for us even goes beyond devices, but for this specific question I would even say that, we want to think about mobility not just one form factor of a mobile device because I think that’s where the ultimate price is.
But that said, we are even year-over-year basis seen increased volume for Lumia, it’s coming at the low end in the entry smartphone market and we are pleased with it. It’s come in many markets we now have over 10% that’s the first market I would sort of say that we need to track country-by-country. And the key places where we are going to differentiate is looking at productivity scenarios or the digital work and life scenario that we can light up on our phone in unique ways.
When I can take my Office Lens App use the camera on the phone take a picture of anything and have it automatically OCR recognized and into OneNote in searchable fashion that’s the unique scenario. What we have done with Surface and PPI shows us the way that there is a lot more we can do with phones by broadly thinking about productivity. So this is not about just a Word or Excel on your phone, it is about thinking about Cortana and Office Lens and those kinds of scenarios in compelling ways. And that’s what at the end of the day is going to drive our differentiation and higher end Lumia phones.
And Keith to answer your specific question, regarding FY16, I think we’ve made the difficult choices to get the cost base to a place where we can deliver, on the exact scenario Satya as outlined, and we do assume that we continue to grow our units through the year and into 2016 in order to get to breakeven.
…
Rick Sherlund – Nomura
Thanks. I’m wondering if you could talk about the Office for a moment. I’m curious whether you think we’ve seen the worst for Office here with the consumer fall off. In Office 365 growth in margins expanding their – just sort of if you can look through the dynamics and give us a sense, do you think you are actually turned the corner there and we may be seeing the worse in terms of Office growth and margins?
Rick, let me just start qualitatively in terms of how I view Office, the category and how it relates to productivity broadly and then I’ll have Amy even specifically talk about margins and what we are seeing in terms of I’m assuming Office renewals is that probably the question. First of all, I believe the category that Office is in, which is productivity broadly for people, the group as well as organization is something that we are investing significantly and seeing significant growth in.
On one end you have new things that we are doing like Cortana. This is for individuals on new form factors like the phones where it’s not about anything that application, but an intelligent agent that knows everything about my calendar, everything about my life and tries to help me with my everyday task.
On the other end, it’s something like Delve which is a completely new tool that’s taking some – what is enterprise search and making it more like the Facebook news feed where it has a graph of all my artifacts, all my people, all my group and uses that graph to give me relevant information and discover. Same thing with Power Q&A and Power BI, it’s a part of Office 365. So we have a pretty expansive view of how we look at Office and what it can do. So that’s the growth strategy and now specifically on Office renewals.
And I would say in general, let me make two comments. In terms of Office on the consumer side between what we sold on prem as well as the Home and Personal we feel quite good with attach continuing to grow and increasing the value prop. So I think that’s to address the consumer portion.
On the commercial portion, we actually saw Office grow as you said this quarter; I think the broader definition that Satya spoke to the Office value prop and we continued to see Office renewed in our enterprise agreement. So in general, I think I feel like we’re in a growth phase for that franchise.
…
Walter Pritchard – Citigroup
Hi, thanks. Satya, I wanted to ask you about two statements that you made, one around responsibly making the market for Windows Phone, just kind of following on Keith’s question here. And that’s a – it’s a really competitive market it feels like ultimately you need to be a very, very meaningful share player in that market to have value for developer to leverage the universal apps that you’re talking about in terms of presentations you’ve given and build in and so forth.
And I’m trying to understand how you can do both of those things once and in terms of responsibly making the market for Windows Phone, it feels difficult given your nearest competitors there are doing things that you might argue or irresponsible in terms of making their market given that they monetize it in different ways?
Yes. One of beauties of universal Windows app is, it aggregates for the first time for us all of our Windows volume. The fact that even what is an app that runs with a mouse and keyboard on the desktop can be in the store and you can have the same app run in the touch-first on a mobile-first way gives developers the entire volume of Windows which is 300 plus million units as opposed to just our 4% share of mobile in the U.S. or 10% in some country.
So that’s really the reason why we are actively making sure that universal Windows apps is available and developers are taking advantage of it, we have great tooling. Because that’s the way we are going to be able to create the broadest opportunity to your very point about developers getting an ROI for building to Windows. For that’s how I think we will do it in a responsible way.
Heather Bellini – Goldman Sachs
Great. Thank you so much for your time. I wanted to ask a question about – Satya your comments about combining the next version of Windows and to one for all devices and just wondering if you look out, I mean you’ve got kind of different SKU segmentations right now, you’ve got enterprise, you’ve got consumer less than 9 inches for free, the offering that you mentioned earlier that you recently announced. How do we think about when you come out with this one version for all devices, how do you see this changing kind of the go-to-market and also kind of a traditional SKU segmentation and pricing that we’ve seen in the past?
Yes. My statement Heather was more to do with just even the engineering approach. The reality is that we actually did not have one Windows; we had multiple Windows operating systems inside of Microsoft. We had one for phone, one for tablets and PCs, one for Xbox, one for even embedded. So we had many, many of these efforts. So now we have one team with the layered architecture that enables us to in fact one for developers bring that collective opportunity with one store, one commerce system, one discoverability mechanism. It also allows us to scale the UI across all screen sizes; it allows us to create this notion of universal Windows apps and being coherent there.
So that’s what more I was referencing and our SKU strategy will remain by segment, we will have multiple SKUs for enterprises, we will have for OEM, we will have for end-users. And so we will – be disclosing and talking about our SKUs as we get further along, but this my statement was more to do with how we are bringing teams together to approach Windows as one ecosystem very differently than we ourselves have done in the past.
Ed Maguire – CLSA
Hi, good afternoon. Satya you made some comments about harmonizing some of the different products across consumer and enterprise and I was curious what your approach is to viewing your different hardware offerings both in phone and with Surface, how you’re go-to-market may change around that and also since you decided to make the operating system for sub 9-inch devices free, how you see the value proposition and your ability to monetize that user base evolving over time?
Yes. The statement I made about bringing together our productivity applications across work and life is to really reflect the notion of dual use because when I think about productivity it doesn’t separate out what I use as a tool for communication with my family and what I use to collaborate at work. So that’s why having this one team that thinks about outlook.com as well as Exchange helps us think about those dual use. Same thing with files and OneDrive and OneDrive for business because we want to have the software have the smart about separating out the state carrying about IT control and data protection while me as an end user get to have the experiences that I want. That’s how we are thinking about harmonizing those digital life and work experiences.
On the hardware side, we would continue to build hardware that fits with these experiences if I understand your question right, which is how will be differentiate our first party hardware, we will build first party hardware that’s creating category, a good example is what we have done with Surface Pro 3. And in other places where we have really changed the Windows business model to encourage a plethora of OEMs to build great hardware and we are seeing that in fact in this holiday season, I think you will see a lot of value notebooks, you will see clamshells. So we will have the full price range of our hardware offering enabled by this new windows business model.
And I think the last part was how will we monetize? Of course, we will again have a combination, we will have our OEM monetization and some of these new business models are about monetizing on the backend with Bing integration as well as our services attached and that’s the reason fundamentally why we have these zero-priced Windows SKUs today.
…
Altera will use Intel Custom Foundry’s 14 nm Tri-Gate (FinFET) process services to produce its new high-end SoC FPGA with 64-bit ARM Cortex-A53 IP
With Stratix® 10 high-end and Arria® 10 mid-range FPGA and SoC FPGA products Altera wants to surge ahead of Xilinx in critical infrastructure—such as wireless remote radio units (RRUs), 100G/400G wireline channel (line) cards and data centers—as well as military, medical and broadcast scenarios by relying on ARM Cortex-A53 IP (Intellectual Property) and Intel Custom Foundry’s 14 nm Tri-Gate (FinFET) process services for Stratix 10, and ARM Cortex-A9 IP and TSMC 20 nm 20SoC process for Arria 10 with OpenCL for FPGAs capability for both. It will also be possible to begin designs with the Arria 10 portfolio of 20 nm FPGA devices, and then take advantage of pin-for-pin design migration pathways from Arria 10 FPGA and SoC products to Stratix 10 FPGA and SoC products as they become available.
This was my conclusion when the news came out that Altera Announces Quad-Core 64-bit ARM Cortex-A53 for Stratix 10 SoCs [press release, Oct 29, 2013] and then I answered three questions for myself, followed by understanding a little bit more deeply two other issues as well:
- Why FPGAs? Why more FPGAs?
- Why SoC FPGAs?
- Why ARM with FPGA on the Intel Tri-Gate (FinFET) process, and why now?
- OpenCL for FPGAs
- Altera SoC FPGAs
For introduction here is Altera Stratix 10 SoC & ARM perspective – ARM TechCon ’13 [ARMflix YouTube channel, Oct 31, 2013]
To shed more light on the direction of breakthrough by Altera, here is additional introductory information from: Arria 10 Device Overview* [Altera, Sept 4, 2013]
*As there is no similar document yet for Stratix 10
Altera’s Arria® FPGAs and SoCs deliver optimal performance and power efficiency in the midrange. By using TSMC’s 20-nm process technology on a high-performance architecture, Arria 10 FPGAs and SoCs deliver higher performance than previous-generation high-end FPGAs while simultaneously reducing power by offering a comprehensive set of power-saving technologies. Altera’s Arria 10 family is reinventing the midrange.
Altera’s Arria 10 SoCs offer a second generation SoC product that both demonstrates a long-term commitment to the SoC product line and extends Altera’s leadership in programmable devices that feature the ARM-based hard processor system (HPS).
Important innovations in Arria 10 devices include:
– Enhanced core architecture delivering 60% higher performance than the previous generation midrange (15% higher performance than previous fastest high-end FPGAs)
– Integrated transceivers with short reach rates up to 28.05 Gbps and backplane capability up to 17.4 Gbps
– Hard PCI Express Gen3 intellectual property (IP) blocks
– Hard memory controllers and PHY up to 2666 Mbps
– Variable precision digital signal processing (DSP) blocks
– Fractional synthesis PLLs
– Up to 40% lower power compared to prior midrange FPGAs and up to 60% lower power compared to prior generation high-end FPGAs due to a comprehensive set of advanced power-saving features
– 2nd generation ARM® Cortex™-A9 hard processor system (HPS) for SoC variants
– Integrated 10GBASE-KR/40GBASE-KR4 Forward Error Correction (FEC)Arria 10 devices are ideally suited for high performance, power-sensitive, midrange applications in such diverse markets as:
– Wireless—for channel and switch cards in remote radio heads and mobile backhaul
– Broadcast—for studio switches, servers and transport, videoconferencing, and pro audio/video
– Wireline—for 40G/100G muxponders and transponders, 100G line cards, bridging, and aggregation
– Compute and Storage—for flash cache, cloud computing servers, and server acceleration
– Medical—for diagnostic scanners and diagnostic imaging
– Military—for missile guidance and control, radar, electronic warfare, and secure communications…
Target Markets for Arria 10 FPGAs and SoCs
Arria 10 devices meet the performance, power, and bandwidth requirements of next generation wireless infrastructure, broadcast, compute and storage, networking, and medical and military equipment.
By providing such a highly integrated device, Arria 10 FPGAs and SoCs significantly reduce BOM cost, form factor, and power consumption. Arria 10 devices allow you to differentiate your product through customization by implementing your intellectual property in both hardware and software.
For these applications, Arria 10 devices integrate both logic functions and processor functions in a highly integrated single device. The integrated ARM-based SoCs provide all the functionality of traditional FPGAs, eliminate the need for a local processor, and increase system performance by taking advantage of the tightly coupled high bandwidth interface between the core fabric and the hard processor system.
For Wireless infrastructure particularly remote radio unit, the industry has standardized onARM-based ASSPs and SoCs for several generations. ARM is widely recognized as the industry leader in low power solutions. At 20 nm, the Dual ARM Cortex MPCore provides the best power efficiency of any GHz class of process. When combined with Altera’s industry leading programmable technology, this provides an ideal platform to address the performance, power, and form factor requirements of wireless remote radio unit and small cell base stations.
For Wireline communication equipment such as access, metro, core,and transmission equipment where the FPGA performs critical functions such as protocol bridging, packet framing, aggregation, and I/O expansion, SoCs now offer all this as well as integrated intelligent controland link management, sometimes referred to as Operations, Administration, and Maintenance (OAM). OAM typically is software that executes when a link is established or fails during operation. The integrated ARM processor can also be used for statistics and error monitoring and minimize system downtime when a link is compromised or oversubscribed. Tight coupling of the processor and the data path (implemented in the core logic) saves time and results in significant savings in terms of operating expenses associated with system downtime and loss of quality of service.
For Compute and storage equipment, flash cache storage, the integrated ARM processor can be used to manage Flash sectors and improve overall life and reliability as well as offload the host processor and provide control for search and hardware acceleration functions for cloud storage equipment. The integrated ARM based HPS can configure the hard PCIe interfaces in PCIe root port configuration and also run link layers for SAS and SATA interfaces.
For Next generation Broadcast equipment, where “4K readiness” is the key technology driver, the integrated ARM processor subsystem eliminates the need for a local GHz class processor, which is commonly used for functions such as audio processing, video compression, video link management, and PCIe root port.
For Military applications, new security features such as Secure Boot, Encryption, and Authentication have been introduced for secure wireless and wireline communications, military radar, military intelligence equipment.
For Test and Medical applications, combining ARM HPS with support for high speed memory devices such as DDR4, and Hybrid Memory Cube (HMC) as well as high speed transceivers and embedded controllers such as PCIe Gen3, Arria 10 SoCs are ideal for next generation test and medical equipment.
Then you can also read The Next-Node Battle Begins – Altera Announces “Generation 10” [EE Journal, June 11, 2013] from I will quote here the following:
For the past three nodes or so, we’ve seen a back-and-forth battle between Altera and Xilinx. Most people think that Altera got the upper hand in 40/45nm products with their Stratix IV family. Two years later, Xilinx struck back hard at 28nm with Virtex-7. Now, it’s time for the “next” generation, and Altera is apparently ready to get the party started. The company has just announced their upcoming “Generation 10” FPGA families – and it looks like this node is gonna be a doozy!
as well as the ARMing a New Generation – Altera Announces Processor Architecture for Gen X [EE Journal, Oct 29, 2013] from which it is wort to quote the following:
Altera is currently in a race with archrival Xilinx, whose first FinFET FPGAs will be riding in on TSMC’s 16nm FinFET process. Which horse is faster? Intel is widely believed to have superior process technology and has already been shipping 22nm FinFET-based devices. Those points go to Intel. TSMC, on the other hand, has vastly more experience as a merchant fab and has announced that they are working closely with Xilinx to accelerate their FinFET program, in a blitz whose marketing name is “FinFAST.”
At this point, therefore, it is unclear who will be shipping first, (and, except for bragging rights between the two companies, probably few people care.) It is likely that we will not see production devices from either company before 2015, so we are definitely in “future” mode here. It is also unclear how the performance attributes of the two companies’ offerings will stack up. Altera has shown more of their hand thus far, and their predictions are impressive – up to four million LUT-4 equivalent 1GHz programmable fabric, 56Gbps SerDes, better power efficiency, tons-o-RAM – and a high-powered processing subsystem in the SoC version. What’s the processing subsystem look like? That’s why we are gathered here today.
There was speculation that the architecture might be other-than-ARM since the manufacturer is none-other-than-Intel. As far as we know, Intel hasn’t historically been too keen on manufacturing competing processor architectures. However, two other, more important market forces are at work in this situation. First, Altera has made a huge commitment to the ARM architecture with their current-generation SoC FPGAs. Getting their customers committed to the ARM/FPGA architecture and then jumping ship and forcing them to migrate after only one generation would be a major inconvenience, and it would be a big black eye for Altera. It would have been very unlikely that Altera would have inked the Intel deal knowing that they couldn’t continue their ARM commitment.
Second, Intel is obviously trying to make a go at it in the merchant fab business. If the company had a hard-and-fast policy of never manufacturing a chip with an ARM architecture on board, they’d be severely limiting their market. While Intel has already been building FPGAs for both Tabula and Achronix, getting Altera in their stable is a whole ‘nuther deal. Putting aside petty concerns about processor architecture is a small price to pay for better street cred in the merchant fab business.
1. Why FPGAs? Why more FPGAs?
As one of the greatest strengths of the FPGA is its ability to perform highly pipelined and complex algorithmic computations on the data brought onchip Altera says that we can do better with explicit parallelism on FPGAs than on GPUs:
The spectrum of software-programmable devices is now evolving significantly. The emphasis is shifting from automatically extracting instruction-level parallelism at run time to explicitly identifying thread-level parallelism at coding time. Highly parallel multicore devices are beginning to emerge with a general trend of containing multiple simpler processors where more of the transistors are dedicated to computation rather than caching and extraction of parallelism. These devices range from multicore CPUs, which commonly have 2, 4, or 8 cores, to GPUs consisting of hundreds of simple cores optimized for data-parallel computation. To achieve high performance on these multicore devices, the programmer must explicitly code their applications in a parallel fashion. Each core must be assigned work in such a way that all cores can cooperate to execute a particular computation. This is also exactly what FPGA designers do to create their high-level system architectures.
(Source: Implementing FPGA Design with the OpenCL Standard
(v. 2.0 Altera whitepaper, November 2012])Field Programmable Gate Arrays
FPGAs are integrated circuits that can be configured repeatedly to perform an infinite number of functions. Low level operations such as bit masking, shifting, and addition are all configurable and can be assembled in any order. FPGAs achieve a high level of programmability by integrating combinations of lookup tables (LUTs), registers, on-chip memories, and arithmetic hardware (for example, digital signal processor (DSP) blocks) through a network of reconfigurable connections to implement computation pipelines. LUTs are responsible for implementing various logic functions. For example, reprogramming a LUT can change an operation from a bitwise AND logic function to a bit-wise XOR logic function.
The key benefit in using FPGAs for algorithm acceleration is that they support wide and heterogeneous pipelines. Each pipeline implemented in the FPGA fabric can be wide and unique. This characteristic is in contrast to many different types of processing units such as symmetric multiprocessors (SMPs), DSPs, and graphics processing units (GPUs). In these types of devices, parallelism is achieved by replicating the same generic computation hardware multiple times. In FPGAs, however, parallelism can be achieved by duplicating only the logic that will be exercised by your algorithm.
A processor implements an instruction set that limits the amount of work that can be performed each clock cycle. For example, most processors do not have a dedicated instruction that can execute the following C code:
E = ((((A + B) ^ C) & D) >> 2;
Without a dedicated instruction for this C code example, a CPU, DSP, or GPU must execute multiple instructions to perform the operation. You can configure an FPGA to perform a sequence of operations that implements the code above in a single clock cycle. An FPGA implementation connects specialized addition hardware with a LUT that performs the bit-wise XOR and AND operations. The device then leverages its programmable connections to perform a right shift by two bits without consuming any hardware resources. The result of this operation can be connected to subsequent operations to form complex pipelines. You may think of an FPGA as a hardware platform that can implement any instruction set that your software algorithm requires.
…
Altera SDK for OpenCL Pipeline Approach
The key difference between the pipeline generated by the Altera Offline Compiler (AOC) and a typical processor pipeline is that the FPGA pipeline is not limited to a statically defined set of pipeline stages or instruction set.
…
The custom pipeline structure provided by the AOC speeds up computation by allowing operations within a large number of threads to occur concurrently.
(Source: Altera SDK for OpenCL Optimization Guide
[for v. 13.0 SP1.0 by Altera, June 2013])
GPU and FPGA Design Methodology
GPUs are programmed using either Nvidia’s proprietary CUDA language, or an open standard OpenCL language. These languages are very similar in capability, with the biggest difference being that CUDA can only be used on Nvidia GPUs.
FPGAs are typically programmed using HDL languages Verilog or VHDL. Neither of these languages is well suited to supporting floating-point designs, although the latest versions do incorporate definition, though not necessarily synthesis, of floating-point numbers. For example, in System Verilog, a short real variable is analogue to an IEEE single (float), and real to an IEEE double.
OpenCL for FPGAs
OpenCL is familiar to GPU programmers. An OpenCL Compiler for FPGAs means that OpenCL code written for AMD or Nvidia GPUs can be compiled onto an FPGA. In addition, an OpenCL Compiler from Altera enables GPU programs to use FPGAs, without the necessity of developing the typical FPGA design skill set.
Using OpenCL with FPGAs offers several key advantages over GPUs. First, GPUs tend to be I/O limited. All input and output data must be passed by the host CPU through the PCI Express® (PCIe®) interface. The resulting delays can stall the GPU processing engines, resulting in lower performance
OpenCL Extensions for FPGAs
FPGAs are well known for their wide variety of high-bandwidth I/O capabilities. These capabilities allow data to stream in and out of the FPGA over Gigabit Ethernet (GbE), Serial RapidIO® (SRIO), or directly from analog-to-digital converters (ADCs) and digital-to-analog converters (DACs). Altera has defined a vendor-specific extension of the OpenCL standard to support streaming operations. …
FPGAs can also offer a much lower processing latency than a GPU, even independent of I/O bottlenecks. It is well known that GPUs must operate on many thousands of threads to perform efficiently, due to the extremely long latencies to and from memory and even between the many processing cores of the GPU. In effect, the GPU must operate many, many tasks to keep the processing cores from stalling as they await data, which results in very long latency for any given task.
The FPGA uses a “coarse-grained parallelism” architecture instead. It creates multiple optimized and parallel datapaths, each of which outputs one result per clock cycle. The number of instances of the datapath depends upon the FPGA resources, but is typically much less than the number of GPU cores. However, each datapath instance has a much higher throughput than a GPU core. The primary benefit of this approach is low latency, a critical performance advantage in many applications.
Another advantage of FPGAs is their much lower power consumption, resulting in dramatically lower GFLOPs/W. FPGA power measurements using development boards show 5-6 GFLOPs/W for algorithms such as Cholesky and QRD, and about 10 GFLOPs/W for simpler algorithms such as FFTs. GPU energy efficiency measurements are much hard to find, but using the GPU performance of 50 GFLOPs for Cholesky and a typical power consumption of 200 W, results in 0.25 GFLOPs/W, which is twenty times more power consumed per useful FLOPs.
(Source: Radar Processing: FPGAs or GPUs? (v. 2.0 Altera whitepaper, May 2013])
Altera also says that the need for ever-increasing bandwidth and flexibility drives the need for a breakthrough in capability:
The increased capabilities in smartphones and other portable devices are the reason for the dramatic leap in system performance that we will see in next-generation FPGAs. The explosion of mobility bandwidth requirements are putting a huge demand on the wireless, wired, and data center infrastructure capabilities. While the number of smartphones is growing at single digit percentage rates, the customers of these devices continue to drive more bandwidth with the ever-increasing smartphone capability. Much of this is due to the increased video content. In 2012, average smartphone data usage grew by 81 percent. Cisco expects mobile traffic to increase 66 percent per year through 2017 and two-thirds of all mobile traffic will be video content. At this time, mobile network speed is expected to increase by seven times and 4G networks to comprise 45 percent of all traffic (1) (see Figure 1).
A brief overview of three infrastructure applications below are examples of why hardware and software developers are looking to FPGAs to address their next-generation products bandwidth, performance, power, and cost goals.
■ Wireless remote radio units
■ 400G wireline channel cards
■ Data centersWireless Remote Radio Units
In the capital-intensive wireless infrastructure market, telecommunications operators desire to provide more bandwidth faster and cheaper. The faster these operators can do cost reductions, the more deployments they can do, the more area they can cover, and the faster they can serve customers—a huge advantage. The product strategy of these companies is to keep the datapath width the same and increase the clock frequency for as many generations as they can. Upcoming remote radio units will look for FPGAs to push close to 500 MHz of core performance for complex functions, such as implementing digital pre-distortion algorithms. This will preserve their investment in their radio architecture and allow them to cover a broader spectrum of radio frequency (RF) bandwidth. In doing so they look to have a better return on investment because less work needs to be done re-architecting a solution. Furthermore, their time-to-market advantage improves by getting these new products out faster. They must also lower their operating costs to drive cost per bit down because revenues per mobile subscriber grow at a far less rate than the data traffic per subscriber. Thus by not widening their datapath, and creating power efficient designs on smaller more power-efficient FPGAs, allows them to achieve this goal.
400G Channel Cards
Another driving force in improving FPGA performance is the need to upgrade the network communications infrastructure. Next-generation 400G versus existing 100G channel cards will dramatically push system capabilities. The bandwidth jump of four times in the next-generation systems is much greater than in previous iterations. Because the market for this is still new, companies cannot risk building ASICs or ASSPs to achieve this goal. Integration of multiple 56 gigabits per second (Gbps) and 28 Gbps transceiver solutions to accommodate this level of bandwidth is needed, but only a part of the solution. More and faster logic to accommodate this higher bandwidth is also required. However since the dimensions of the chassis do not change, the power envelope is limited. The network infrastructure cannot tolerate solutions where power increases at a linear rate with bandwidth capability. For packet processing and traffic management applications at 400G bandwidth at 600 million packets per second, scaling the data path width and frequency can relieve the data path processing function but cannot scale for control path processing such as scheduling. Therefore high performance in all aspects of device capability is required: processing, memory interfacing, IO interfaces, and others. FPGAs remain the most attractive solution, but companies will need investments in higher performance per watt architectures, transceivers, and process technology to address this large leap in capabilities and challenges.
Data Centers
All the data and video that are being pushed and downloaded from these new wireless deployments and transported through the new 400G packet processing infrastructure also needs to be stored and processed. Computations per watt and computations per dollar is a key metric in data centers. FPGA’s are increasingly used in the data center for data access, algorithm, and networking acceleration. Data center servers are bottlenecked getting access to data. The latest processors have more and more cores, but the bandwidth to external memory and data is not keeping pace with the increase in computing power. Many of these servers are running at average utilization rates and are well under peak processing power. These servers are good candidates for FPGA acceleration. Hardware acceleration through FPGAs becomes an attractive alternative to replacing these processors by focusing on the performance bottlenecks that software on processors cannot overcome.
Other applications are also looking to FPGAs to support their increased bandwidth requirements, such as video content providers moving to 4K video, cloud computing, and intelligence applications in defense. These applications face similar issues. (Source: Expect a Breakthrough Advantage in Next-Generation FPGAs (v. 1.0 Altera whitepaper, June 2013])
2. Why SoC FPGAs?
Altera’s Vision of Silicon Convergence: system solutions by merging coarse and fine grained programmable hardware [IEEE Computer Society Santa Clara Valley YouTube channel, recorded on Sept 10, 2012, published on June 10, 2013]
What Is a PLD?
- A programmable logic device (PLD) is a type of semiconductor
- Most semiconductors can be programmed only once to perform a specific function
- PLDs are reprogrammable—functions can be changed or enhanced during development or after manufacturing
Flexibility Makes PLDs Lower Risk and Faster to
Design Than Other Types of Semiconductors
3. Why ARM with FPGA on the Intel Tri-Gate (FinFET) process, and why now?
Altera Announces Quad-Core 64-bit ARM Cortex-A53 for Stratix 10 SoCs [press release, Oct 29, 2013]
Manufactured on Intel’s 14 nm Tri-Gate Process, Altera Stratix® 10 SoCs Will Deliver Industry’s Most Versatile Heterogeneous Computing Platform
Altera Corporation (NASDAQ: ALTR) today announced that its Stratix 10 SoC devices, manufactured on Intel’s 14 nm Tri-Gate process, will incorporate a high-performance, quad-core 64-bit ARM Cortex™-A53 processor system, complementing the device’s floating-point digital signal processing (DSP) blocks and high-performance FPGA fabric. Coupled with Altera’s advanced system-level design tools, including OpenCL, this versatile heterogeneous computing platform will offer exceptional adaptability, performance, power efficiency and design productivity for a broad range of applications, including data center computing acceleration, radar systems and communications infrastructure.
The ARM Cortex-A53 processor, the first 64-bit processor used on a SoC FPGA, is an ideal fit for use in Stratix 10 SoCs due to its performance, power efficiency, data throughput and advanced features. The Cortex-A53 is among the most power efficient of ARM’s application-class processors, and when delivered on the 14 nm Tri-Gate process will achieve more than six times more data throughput compared to today’s highest performing SoC FPGAs. The Cortex-A53 also delivers important features, such as virtualization support, 256TB memory reach and error correction code (ECC) on L1 and L2 caches. Furthermore, the Cortex-A53 core can run in 32-bit mode, which will run Cortex-A9 operating systems and code unmodified, allowing a smooth upgrade path from Altera’s 28 nm and 20 nm SoC FPGAs.
“ARM is pleased to see Altera adopting the lowest power 64-bit architecture as an ideal complement to DSP and FPGA processing elements to create a cutting-edge heterogeneous computing platform,” said Tom Cronk, executive vice president and general manager, Processor Division, ARM. “The Cortex-A53 processor delivers industry-leading power efficiency and outstanding performance levels, and it is supported by the ARM ecosystem and its innovative software community.”
Leveraging Intel’s 14 nm Tri-Gate process and an enhanced high-performance architecture, Altera Stratix 10 SoCs will have a programmable-logic performance level of more than 1GHz; two times the core performance of current high-end 28 nm FPGAs.
“High-end networking and communications infrastructure are rapidly migrating toward heterogeneous computing architectures to achieve maximum system performance and power efficiency,” said Linley Gwennap, principal analyst at The Linley Group, a leading embedded research firm. “What Altera is doing with its Stratix 10 SoC, both in terms of silicon convergence and high-level design tool support, puts the company at the forefront of delivering heterogeneous computing platforms and positions them well to capitalize on myriad opportunities.”
By standardizing on ARM processors across its three-generation SoC portfolio, Altera will offer software compatibility and a common ARM ecosystem of tools and operating system support. Embedded developers will be able to accelerate debug cycles with Altera’s SoC Embedded Design Suite (EDS) featuring the ARM Development Studio 5 (DS-5™) Altera® Edition toolkit, the industry’s only FPGA-adaptive debug tool, as well as use Altera’s software development kit (SDK) for OpenCL to create heterogeneous implementations using the OpenCL high-level design language.
“With Stratix 10 SoCs, designers will have a versatile and powerful heterogeneous compute platform enabling them to innovate and get to market faster,” said Danny Biran, senior vice president, corporate strategy and marketing at Altera. “This will be very exciting for customers as converged silicon continues to be the best solution for complex, high-performance applications.”
About Altera
Altera® programmable solutions enable designers of electronic systems to rapidly and cost effectively innovate, differentiate and win in their markets. Altera offers FPGAs, SoCs, CPLDs, ASICs and complementary technologies, such as power management, to provide high-value solutions to customers worldwide. Follow Altera viaFacebook, Twitter, LinkedIn, Google+ and RSS, andsubscribe to product update emails and newsletters. altera.com
Altera to Build Next-Generation, High-Performance FPGAs on Intel’s 14 nm Tri-Gate Technology [alteracorp YouTube channel, March 11, 2013]
From: Intel takes big step in chip foundry business [Reuters, Feb 25, 2013]
Altera Chief Executive John Daane told Reuters in a phone interview that Altera, which depends on communications infrastructure for about half of its business, is the only major programmable chipmaker that will have access to Intel’s plants.
“We are essentially getting access like an extra division of Intel. As soon as they’re making the technology available to their various groups to do design work, we’re getting the same,” he said.
Daane said Intel’s manufacturing technology will give Altera’s chips a several-year advantage against Xilinx, its main competitor in programmable chips. He said Altera would continue to make other chips with TSMC, its long-time foundry.
Altera to Build Next-Generation, High-Performance FPGAs on Intel’s 14 nm Tri-Gate Technology [press release, Feb 25, 2013]
Altera Corporation and Intel Corporation today announced that the companies have entered into an agreement for the future manufacture of Altera FPGAs on Intel’s 14 nm tri-gate transistor technology. These next-generation products, which target ultra high-performance systems for military, wireline communications, cloud networking, and compute and storage applications, will enable breakthrough levels of performance and power efficiencies not otherwise possible.
“Altera’s FPGAs using Intel 14 nm technology will enable customers to design with the most advanced, highest-performing FPGAs in the industry,” said John Daane, president, CEO and chairman of Altera. “In addition, Altera gains a tremendous competitive advantage at the high end in that we are the only major FPGA company with access to this technology.”
Altera’s next-generation products will now include 14 nm, in addition to previously announced 20 nm technologies, extending the company’s tailored product portfolio that meets myriad customer needs for performance, bandwidth and power efficiency across diverse end applications.
“We look forward to collaborating with Altera on manufacturing leading-edge FPGAs, leveraging Intel’s leadership in process technology,” said Brian Krzanich, chief operating officer, Intel. “Next-generation products from Altera require the highest performance and most power-efficient technology available, and Intel is well positioned to provide the most advanced offerings.”
Adding this world-class manufacturer to Altera’s strong foundation of leading-edge suppliers and partners furthers the company’s ability to deliver on the promise of silicon convergence; to integrate hardware and software programmability, microprocessors, digital signal processing, and ASIC capability into a single device; and deliver a more flexible and economical alternative to traditional ASICs and ASSPs.
Altera claims that only Intel’s 14 nm Tri-Gate Process offers a second generation of proven production technology:
Transistor Design Background
In 1947 the first transistor, a germanium ‘point-contact’ structure, was demonstrated at Bell Laboratories. Silicon was first used to produce bipolar transistors in 1954, but it was not until 1960 that the first silicon metal oxide semiconductor field-effect transistor (MOSFET) was built. The earliest MOSFETs were 2D planar devices with current flowing along the surface of the silicon under the gate. The basic structure of MOSFET devices has remained substantially unchanged for over 50 years.
Since the prediction or proclamation of Moore’s Law in 1965, many additional enhancements and improvements have been made to the manufacture and optimization of MOSFET technology in order to enshrine Moore’s Law in the vocabulary and product planning cycles of the semiconductor industry. In the last 10 years, the continued improvement in MOSFET performance and power has been achieved by breakthroughs in strained silicon, and High-K metal gate technology.
It was not until the publication of a paper by Digh Hisamoto and a team of other researchers at Hitachi Central Research Laboratory in 1991 that the potential for 3-D, or ‘wraparound’ gate transistor technology, to enhance MOSFET performance and eliminate short channel effects, was recognized. This paper called the proposed 3-D structure ‘depleted lean-channel transistor’, or DELTA(1). In 1997 the Defense Advanced Research Projects Agency (DARPA) awarded a contract to a research group at the University of California, Berkeley, to develop a deep sub-micron transistor based on the DELTA concept. One of the earliest publications resulting from this research in 1999 dubbed the device a ‘FinFET’ for the fin-like structure at the center of the transistor geometry(2).
Important Turning Point in Transistor Technology
Continued optimization and manufacturability studies on 3-D transistor structures continued at research and development organizations in leading semiconductor companies. Some of the process and patent development has been published and publicly shared, and some development remained in corporate labs.
The research investment interests of the semiconductor industry are driven by the International Technology Roadmap for Semiconductors (ITRS), which is coordinated and published by a consortium of manufacturers, suppliers, and research institutes. The ITRS defines transistor technology requirements to achieve continued improvement in performance, power, and density along with options which should be explored to achieve the goals. The ITRS and its public documentation captures conclusions and recommendations regarding manufacturing capabilities like strained silicon and High-K metal gate, and now the use of 3-D transistor technologies to maintain the benefits of Moore’s law. Based on documents produced by the ITRS and an examination of academic papers and patent filings, research into 3-D transistor technologies has grown dramatically in the last decade.
Adoption and Research
Two important pronouncements occurred in the last two years that have propelled the 3-D transistor structure into the industry spotlight, and into a permanent place in the technology story of MOSFET transistors.
The first announcement was by Intel Corporation on 4th of May, 2011, about their Tri-Gate transistor design that had been selected for the design and manufacture of their 22 nm semiconductor products. This was preceded by a decade of research and development taking advantage of the work of Hisamoto and others in FinFET development and optimization. It represented both a solid acknowledgment of the feasibility and cost-effectiveness of the the Tri-Gate transistor structure in semiconductor production, as well as a continued declaration of leadership by Intel in semiconductor technology.
The second announcement was the publication of ITRS technology roadmaps, with contributions from many other semiconductor manufacturing companies that identified 3-D transistor technology as the primary enabler of all incremental semiconductor improvement beyond the 20 nm or 22 nm design node.
…
Intel’s Leadership in Transistor Technologies
In several public forums, including the Intel Developer’s Forums and investor’s conferences, Intel identifies where they have demonstrated technology leadership in a variety of advances that have sustained the pace of Moore’s Law. As shown in Figure 3, Intel has identified the number of years of production leadership they have achieved in bringing strained silicon and High-K metal gate technology to full production. In the case of 3-D Tri-Gate transistor technology, Intel estimates a lead of up to four years based on their production rollout of Tri-Gate technology at 22 nm in 2011.
According to former Intel CEO, Paul Otellini in their 16 April 2013 Earnings Call(8):
“In the first quarter [of 2013], we shipped our 100 millionth 22 nanometer [Tri-Gate] processor, using our revolutionary 3-D transistor technology, while the rest of the industry works to ship its first unit.”
Another leadership advantage that will be held by Intel in their rollout of 14 nm technology can be traced to their very public ‘Tick-Tock’ strategy in process and microarchitecture introduction. A ‘tick’ cycle of product introduction relies on the implementation of microarchitecture changes in their CPU products, followed by a ‘tock’ cycle of semiconductor process manufacturing geometry shrink. Intel is firmly committed to a full process shrink in their move from 22 nm to 14 nm; comparable semiconductor technology processes in development at other manufacturers have been less clear whether their process roadmaps include the benefits of a process shrink.
(Source: The Breakthrough Advantage for FPGAs with Tri-Gate Technology (v. 1.0 Altera whitepaper, June 2013])
Altera says beginning with 14 nm Tri-Gate technology, the highest performance FPGAs will simply be the ones built on demonstrably superior transistor technology:
Accessing the Benefits of Tri-Gate Technology Through Altera FPGAs
Taking advantage of the significant benefits of Intel’s Tri-Gate technology is only possible for users of Altera® high-density and high-performance FPGAs on the 14 nm technology process. This is the result of an exclusive manufacturing partnership between the two companies referenced in the introduction to this paper.
The substantial advantages of Tri-Gate silicon technologies will allow Altera to deliver previously unimaginable performance in FPGA and SoC products. This will include a historic doubling of core performance as compared to other high-end FPGAs, bringing FPGAs to the Gigahertz performance level. Overall active and static power numbers will reduce by 70 percent through a combination of process, architecture, and software advances.
Although the details and schedules of the 14 nm manufacturing process are not yet publicly available from Intel Corporation, Altera users can begin designs today that take advantage of the significant performance and power efficiency benefits of Tri-Gate technology in FPGAs. This is possible by beginning designs with the Arria® 10 portfolio of 20 nm FPGA devices. Users can then take advantage of pin-for-pin design migration pathways from Arria 10 FPGA and SoC products to Stratix® 10 FPGA and SoC products as they become available.
This allows you, as an FPGA user and system architect, to begin designing products that can accommodate both the Arria 10 and Stratix 10 product families with minimal changes, modifications, and reengineering. This will allow you to get products to market with the highest performance and lowest power FPGAs that leverage 20 nm process technology and power reduction techniques, then advance these same products to the previously unimaginable performance and power efficiency of Intel’s 14 nm Tri-Gate manufacturing process.
(Source: The Breakthrough Advantage for FPGAs with Tri-Gate Technology (v. 1.0 Altera whitepaper, June 2013])
Altera Announces Breakthrough Advantages with Generation 10 [press release, June 10, 2013]
- Stratix 10 FPGAs and SoCs leverage Intel’s 14 nm Tri-Gate process and an enhanced architecture to deliver core performance two times higher than current high-end FPGAs, while enabling up to 70 percent power savings.
- Arria 10 FPGAs and SoCs reinvent the midrange by simultaneously surpassing high-end FPGAs in performance while delivering 40 percent lower power than today’s midrange devices.
Altera Corporation (NASDAQ: ALTR) today introduced its Generation 10 FPGAs and SoCs, offering system developers breakthrough levels of performance and power efficiencies. Generation 10 devices are optimized based on process technology and architecture to deliver the industry’s highest performance and highest levels of system integration at the lowest power. Initial Generation 10 families include Arria® 10 and Stratix® 10 FPGAs and SoCs with embedded processors. Generation 10 devices leverage the most advanced process technologies in the industry, including Intel’s 14-nm Tri-Gate process and TSMC’s 20 nm process. Early access customers are currently using the Quartus® II software for Generation 10 product development.
“Our Generation 10 products will strengthen the penetration of programmable logic into new markets and applications and further accelerate the implementation of FPGAs into systems traditionally served by ASSPs and ASICs,” said Patrick Dorsey, senior director of product marketing at Altera. “The optimizations we made in our Generation 10 devices allow customers to develop highly customized solutions that dramatically increase system performance and system integration while lowering operating expenses.”
Delivering the Unimaginable with Stratix 10 FPGAs and SoCs
Stratix 10 FPGAs and SoCs are designed to enable the most advanced, highest performance applications in the communications, military, broadcast and compute and storage markets, while slashing system power. Leveraging Intel’s 14 nm Tri-Gate process and an enhanced high-performance architecture, Stratix 10 FPGAs and SoCs have an operating frequency over one gigahertz, 2X the core performance of current high-end 28 nm FPGAs. For high-performance systems that have the most strict power budgets, Stratix 10 devices allow customers to achieve up to a 70 percent reduction in power consumption at performance levels equivalent to the previous generation.
Altera is announcing the technology details of Stratix 10 FPGAs and SoCs today as part of the Generation 10 portfolio introduction, and will disclose more details on the product at a later date. Stratix 10 FPGAs and SoCs provide the industry’s highest performance and highest levels of system integration, including:
- More than four million logic elements (LEs) on a single die
- 56-Gbps transceivers
- More than 10-TeraFLOPs single-precision digital signal processing
- A third-generation ultra-high-performance processor system
- Multi-die 3D solutions capable of integrating SRAM, DRAM and ASICs
Reinventing the Midrange with Arria 10 FPGAs and SoCs
Arria 10 FPGAs and SoCs are the first device families to roll out as part of the Generation 10 portfolio. The device family sets a new bar for midrange programmable devices, delivering both the performance and capabilities of current high-end FPGAs at the lowest midrange power. Leveraging an enhanced architecture that is optimized for TSMC’s 20 nm process, Arria 10 FPGAs and SoCs deliver higher performance at up to 40 percent lower power compared to the previous device family.
Arria 10 devices offer more features and capabilities than today’s current high-end FPGAs, at 15 percent higher performance. Reflecting the trend toward silicon convergence, Arria 10 FPGAs and SoCs offer the highest degree of system integration available in midrange devices, including 1.15 million LEs, integrated hard intellectual property and a second-generation processor system that features a 1.5 GHz dual-core ARM® Cortex™-A9 processor. Arria 10 FPGAs and SoCs also provide 4X greater bandwidth compared to the current generation, including 28-Gbps transceivers, and 3X higher system performance, including 2666 Mbps DDR4 support and up to 15-Gbps Hybrid Memory Cube support.
Development Suite Delivers Breakthrough Productivity to Generation 10
Generation 10 devices are supported by Altera’s Quartus II development software and tools for higher level design flows that include a software development kit for OpenCL™, a SoC Embedded Design Suite and DSP Builder tool. This leading-edge development tool suite enables design teams to maximize productivity while making it easier for new design teams to adopt Generation 10 FPGAs and SoCs in their next-generation systems. The Quartus II software will continue to deliver the industry’s fastest compile times by providing Generation 10 FPGAs and SoCs an 8X improvement in compile times versus the previous generation. The substantial reduction in compile times is the result of leading-edge software algorithms that take advantage of modern multi-core computing technologies.
Availability
Early access customers are currently using the Quartus II software for development of Arria 10 FPGA and SoCs. Initial samples of Arria 10 devices will be available in early 2014. Altera will have 14 nm Stratix 10 FPGA test chips in 2013 and Quartus II software support for Stratix 10 FPGAs and SoCs in 2014. For more information, visit www.altera.com/gen10, or contact your local Altera sales representative.
Altera and TSMC Continue Long-Term Partnership [press release, Feb 25, 2013]
Altera Corporation (NASDAQ: ALTR) and TSMC (TWSE: 2330, NYSE: TSM) today reaffirmed their commitment to a long-term partnership to set new milestones in FPGA innovation. TSMC is Altera’s primary foundry, supplying a wide array of processes to fulfill Altera’s product portfolio, including soon-to-be released 20 nm products, existing mainstream products, and long-lived legacy components.
Altera is fully engaged with TSMC on developing products based on next-generation process technologies. Altera’s next major product family leverages TSMC’s cost-effective 20SoC process for optimal power and performance and will include several significant product and technology innovations for both companies. Altera will continue to leverage future TSMC process technologies in its tailored product portfolio for performance, bandwidth, and power efficiency needs across diverse end applications.
“Over the course of our 20-year collaboration, Altera and TSMC have achieved many industry milestones that have greatly benefitted both companies,” said John Daane, president, CEO and chairman of Altera. “TSMC remains an important part of our future product development. We look forward to continuing our close partnership to jointly develop technologies for next-generation products.”
Morris Chang, TSMC’s chairman and CEO added,”The history of collaboration between Altera and TSMC has exemplified the way fabless and foundry have nurtured each other to become a powerful force in the semiconductor industry. TSMC would not be where it is today without customers like Altera, and I firmly believe this partnership will continue to flourish.”
Altera Demonstrates Industry’s First 32-Gbps Transceiver with Leading-Edge 20 nm Device [press release, April 8, 2013]
Demonstration Highlights Latest Success in Altera’s 20 nm FPGA Early Access Program
San Jose, Calif., April 8, 2013– Altera Corporation (NASDAQ: ALTR) today announced the company achieved another significant milestone in transceiver technology by demonstrating the industry’s first programmable device with 32-Gbps transceiver capabilities. The demonstration uses a 20 nm device based on TSMC’s 20SoC process technology. This achievement validates the performance capabilities of 20 nm silicon and is a positive indicator to the more than 500 customers in Altera’s early access program who are looking to use next-generation Altera devices in the development of performance demanding, bandwidth-centric applications. A demonstration video showing the industry’s first operational 20 nm transceiver technology operating at 32 Gbps is available for viewing on Altera’s website at www.altera.com/32gbps-20nm.
Demonstrating 32-Gbps transceiver data rates provides Altera insight into how high-performance transceiver designs behave on TSMC’s 20SoC process. The transceiver technology Altera is demonstrating today will be integrated into its 20 nm FPGA products, fabricated on TSMC’s 20SoC process. These devices enable customers to design next-generation serial links with the lowest power consumption, fastest timing closure and the highest quality signal integrity. Altera has a proven track record in integrating leading-edge transceiver technology into its devices. Altera is the only company today shipping production 28 nm FPGAs with monolithically integrated low-power transceivers operating at 28 Gbps. Being the first FPGA vendor to reach the 32-Gbps milestone in 20 nm silicon further extends Altera’s leadership in transceiver technology.
The demonstration video on Altera’s web site shows 20 nm transceivers operating at 32 Gbps with just over nine picoseconds of total jitter and extremely low random jitter of 240 femtoseconds. The results show good margin to key industry specifications requited for next-generation 100G systems.
“Today’s news represents a significant milestone for the industry and for the transceiver development team at Altera,” said Vince Hu, vice president of product and corporate marketing at Altera. “These 20 nm devices contain the key IP components that will be included in our next-generation FPGAs and validating them now provides us confidence we will deliver to the market 20 nm FPGAs on schedule.”
Altera’s next-generation transceiver innovations enable system developers to support the rapidly increasing amount of data that is being transmitted through the world’s networks. The transceivers in Altera’s next-generation devices will drive more bandwidth with lower power per channel versus the previous nodes and will support increasing port density by interfacing directly to 100G CPF2 optical modules.
Altera and Micron Lead Industry with FPGA and Hybrid Memory Cube Interoperability [joint press release, Sept 4, 2013]
Altera Corporation (NASDAQ: ALTR) and Micron Technology, Inc.(NASDAQ: MU) (“Micron”) today announced they have jointly demonstrated successful interoperability between Altera Stratix® V FPGAs and Micron’s Hybrid Memory Cube (HMC). This technology achievement enables system designers to evaluate today the benefits of HMC with FPGAs and SoCs for next-generation communications and high-performance computing designs. The demonstration provides an early proof point that production support of HMC will be delivered with Altera’s Generation 10 portfolio, in alignment with market timing, and includes both Stratix 10 and Arria 10 FPGAs and SoCs.
HMC has been recognized by industry leaders and influencers as the long-awaited answer to address the limitations imposed by conventional memory technology, and provides ultra-high system performance with significantly lower power-per-bit. HMC delivers up to 15 times the bandwidth of a DDR3 module and uses 70 percent less energy and 90 percent less space than existing technologies. HMC’s abstracted memory allows designers to devote more time leveraging HMC’s revolutionary features and performance and less time navigating the multitude of memory parameters required to implement basic functions. It also manages error correction, resiliency, refresh, and other parameters exacerbated by memory process variation. Micron expects to begin sampling HMC later this year with volume production ramping in 2014.
“As one of the founding developers of the HMC Consortium, Altera’s support for and involvement with HMC has been invaluable,” said Brian Shirley, vice president of DRAM solutions for Micron Technology. “The combination of Altera FPGAs with Micron’s HMC solution will help customers leverage the technology’s performance and efficiency in a wide range of next generation networking and computing applications.”
Altera’s 28 nm Stratix V FPGAs are an ideal demonstration of HMC technology since they are the highest performance FPGAs in the industry with a two speed-grade advantage over the nearest competitor. This performance enables the FPGA to leverage the full bandwidth, efficiency and power benefits of HMC by using a full 16 transceiver HMC link.
“By demonstrating Stratix V and HMC working together now, we are enabling our customers to leverage their current development with Stratix V FPGAs and prepare for production deployment in Altera’s Generation 10 devices, knowing they will have proven HMC support,” said Danny Biran, senior vice president of marketing and corporate strategy at Altera. “The partnership between Altera and Micron to deliver this capability puts our customers at the forefront of innovation.”
Altera’s Generation 10 Devices Deliver Performance
Arria 10 FPGAs and SoCs are the first device families in the Generation 10 portfolio and will be the first devices to support HMC technology in volume production. Leveraging an enhanced architecture optimized for TSMC’s 20 nm process, Arria 10 FPGAs and SoCs will use HMC to extend the benefits by providing both 15 percent higher core performance than today’s highest performance Stratix V FPGAs and up to 40 percent lower power compared to the lowest power Arria V midrange FPGAs. Arria 10 FPGAs and SoCs will offer up to 96 transceiver channels, enabling customers to take full advantage of the bandwidth that HMC has to offer.
Stratix 10 FPGAs and SoCs will enable the most advanced, highest performance applications across communications, military, broadcast and compute and storage markets. These high-performance applications often require the highest memory bandwidth, which drives the need for an HMC-ready architecture. Leveraging Intel’s 14 nm Tri-Gate process and an enhanced high-performance architecture that integrates with HMC technology, Stratix 10 FPGAs and SoCs will enable system solutions with an operating frequency over one gigahertz, and two times the core performance of current high-end 28 nm FPGAs. Stratix 10 devices will also allow customers to achieve up to a 70 percent reduction in power consumption at performance levels equivalent to the previous generation.
4. OpenCL for FPGAs
Altera SDK for OpenCL is First in Industry to Achieve Khronos Conformance for FPGAs [press release, Oct 16, 2013]
Altera Passes OpenCL Conformance with High-Performance Stratix V FPGA and Demonstrates SDK for OpenCL on ARM-based Cyclone V SoCs
San Jose, Calif., October 16, 2013—Altera Corporation (NASDAQ: ALTR) today announced its SDK for OpenCL is conformant to the OpenCL 1.0 standard and is now included on the Khronos Group list of OpenCL conformant products. Altera is the only company to offer an FPGA-optimized OpenCL solution, allowing software developers to harness the massively parallel architecture of an FPGA for system acceleration. Altera will demonstrate its OpenCL solutions at the 2013 Linley Processor Conference, being held October 16-17 in Santa Clara, Calif.
Achieving conformance allows Altera to provide a validated cross-platform programming environment that can be used to dramatically accelerate algorithms at significantly lower power versus alternative computer hardware architectures. To become conformant, Altera successfully completed more than 8500 conformance tests using its SDK for OpenCL, targeting a high-performance Stratix® V FPGA. The tests involved continuously running a Stratix V FPGA accelerator card in a server farm resulting in zero errors.
“Our continued investment in OpenCL is enabling Altera to drive the industry toward using FPGAs for acceleration of computationally-intensive applications,” said Alex Grbic, director of software, IP and DSP marketing at Altera. “Our SDK for OpenCL is used by some of the world’s leading developers of high-performance computing systems. These developers require Khronos group OpenCL conformance and Altera is the only FPGA vendor to achieve it, proving the readiness of our solution.”
Software developers can easily take advantage of the high-performance, low-power that FPGAs offer. Altera’s SDK for OpenCL provides an industry-standard open source programming interface and Altera’s Preferred Board Partner Program for OpenCL provides off-the-shelf FPGA boards that are optimized for Altera devices. A list of preferred board partners, as well as a variety of design examples that demonstrate the advantages of using FPGAs in high-performance systems, can be found at www.altera.com/opencl.
OpenCL Ray Tracer Demonstration Targeting Single-chip SoCs
In addition to support for its high-performance Stratix V FPGAs, Altera developed its SDK for OpenCL to support its low-power, low-cost Cyclone® V SoCs, which integrates an ARM® Cortex®-A9 processor into a 28 nm FPGA. Altera recently used its SDK for OpenCL to develop and demonstrate a complete heterogeneous system using a Cyclone V SoC. The demonstration shows how a ray tracing algorithm used to render 3D graphics can be accelerated using the Altera SDK for OpenCL and a Cyclone V SoC – achieving a speed up of 40X in comparison to running the same algorithm purely on a discrete ARM processor system. For software developers unfamiliar with hardware design languages, no hardware expertise is required to implement the OpenCL kernels.
Altera SDK for OpenCL at Linley Processor Conference
Altera will demonstrate its OpenCL solutions at the 2013 Linley Tech Processor Conference, being held October 16-17 in Santa Clara, Calif. Altera’s participation includes a presentation titled “Implementing Deep Packet Inspection Using OpenCL Channels” that will show how to express a DPI application using OpenCL with Altera FPGAs. Altera will also demonstrate its SDK for OpenCL solutions to attendees.
Pricing and Availability
The Altera SDK for OpenCL is currently available for download on Altera’s website or through the purchase of an Altera Preferred Partner OpenCL board. The annual software subscription for the SDK for OpenCL is $995. For additional information please visit the OpenCL section on Altera’s website.
LEAP 2013 : Developing High-Performance Low-Power Solutions using FPGAs and OpenCL by Craig Davis — Altera Corporation [LEAPconf YouTube channel, recorded on May 21, 2013, published on Sept 12, 2013]
From presentation slides (PDF) I will copy here the following ones:
Programming Models
FPGA programming model: RTL
Involves state machines, datapaths, arbitration, buffering, and others
Processor programming model: C/C++
Typically sequential, involves subroutines and functions
Need a programming model that represents a heterogeneous system (CPU + FPGA)
A processor with hardware accelerators
A configurable multicore device
The goal
An ideal single hardware and software design environment
More information: Implementing FPGA Design with the OpenCL Standard (v. 2.0 Altera whitepaper, November 2012]
Unified Heterogeneous Programmability of OpenCL [alteracorp YouTube channel, Nov 5, 2012]
Altera Opens the World of FPGAs to Software Programmers with Broad Availability of SDK and Off-the-Shelf Boards for OpenCL [press release, May 6, 2013]
Altera SDK for OpenCL Combined with an Ecosystem of Development Boards Delivers Power-efficient, High-performance Solution for Heterogeneous Computing
Altera Corporation (NASDAQ: ALTR) today announced the broad availability of its SDK for OpenCL™ and supported third-party production boards. Availability of the SDK for OpenCL enables software programmers to access the high-performance capabilities of programmable logic devices. Also part of today’s news, Altera announced a Preferred Board Partner Program, allowing third-party board vendors to work closely with Altera to design optimized production boards based on Altera’s programmable devices. The availability of supported third-party boards through the Preferred Board Partner Program and an SDK for OpenCL enables software programmers to easily target high-performance FPGAs using a high-level language.
Altera’s SDK for OpenCL allows software programmers to take their OpenCL code and easily exploit the massively parallel architecture of an FPGA. Software programmers targeting FPGAs achieve higher performance at significantly lower power compared to alternative hardware architectures.
“Because FPGAs enable parallel processing, they are critical for specialized server workloads that demand real-time performance. We are pleased that our clients are now able to take full advantage of this technology on Power Systems using Altera’s SDK for OpenCL,” said Robert L. Swann, vice president, IBM Power Systems. “With this standards-based approach, our clients can leverage a vibrant ecosystem of commercial and research contributions to accelerate emerging compute intensive workloads.”
The SDK for OpenCL is designed to increase system performance in highly data-parallel computing applications featured in financial, military, broadcast, medical and a variety of other markets. Altera’s OpenCL solutions are supported by a robust ecosystem consisting of board partners, design partners, software tools and university collaboration. Altera and its partners provide the tools, hardware, libraries, reference designs and design resources necessary for developers to implement their OpenCL designs into FPGAs and reduce time-to-market.
The Altera Preferred Board Partner Program for OpenCL ensures third-party production boards are optimized for current Altera device architectures. Initial preferred board partners included in the program are BittWare, Nallatech and PLDA, with additional board partners to be added in the future.
“For years, Altera and BittWare have partnered to deliver timely high-end signal processing board-level solutions that significantly reduce technology risk for our mutual customers,” said Darren Taylor, senior vice president of sales and marketing at BittWare. “Leveraging the latest hardware technology from Altera, which now includes an SDK for OpenCL, we are able to dramatically reduce the complexity for applications in the computing, financial and military markets.”
“An OpenCL implementation provides an ideal fit for Nallatech’s hardware-accelerated computing solutions,” said Allan Cantle, president and founder of Nallatech. “We simplify the deployment of FPGAs in heterogeneous platforms via direct purchase of our cards or pre-integrated in leading vendors’ high density servers and blades. Customers developing high-performance computing applications using Altera’s SDK for OpenCL will benefit from a dramatic increase in performance per watt, per dollar over traditional computing architectures.”
“PLDA has a successful track record of supporting Altera’s customers with their high-performance applications,” said Stephane Hauradou, vice president and CTO of PLDA. “The SDK for OpenCL will open up a significantly broader group of software developers who can now fully leverage Altera’s leading-edge solutions.”
Pricing and Availability
The Altera SDK for OpenCL is currently available for download on Altera’s website. The annual software subscription for the SDK for OpenCL is $995 for a node-locked PC license. For additional information about the Altera Preferred Board Partner Program for OpenCL and its partner members, or to see a list of all supported boards and links to purchase, visit the OpenCL section on Altera’s website.
Altera Announces Industry’s First FPGA Support for OpenCL – Eases the Adoption of FPGAs for Accelerating Heterogeneous Systems [press release, Nov 5, 2012]
Software Development Kit for OpenCL Enables Developers to Take Advantage of the Performance and Power-efficiencies of FPGAs
Altera Corporation (Nasdaq: ALTR) today announced the FPGA industry’s first Software Development Kit (SDK) for OpenCL™ (Open Computing Language) which combines the massively parallel architecture of an FPGA with the OpenCL parallel programming model. The SDK allows system developers and programmers familiar with C to quickly and easily develop high-performance, power-efficient FPGA-based applications in a high-level language. The Altera SDK for OpenCL enables FPGAs to work in concert with the host processor to accelerate parallel computation, at a fraction of the power compared to hardware alternatives. Altera will demonstrate the performance and productivity benefits of OpenCL for FPGAs at SuperComputing 2012 in booth #430.
“The industry’s approach for boosting system performance has evolved over time from increasing frequency in single-core CPUs, to using multi-core CPUs, to using parallel processor arrays,” said Vince Hu, vice president of product and corporate marketing at Altera. “This evolution leads us to today’s modern FPGAs, which are fine-grained, massively parallel digital logic arrays architected to execute computations in parallel. Our SDK for OpenCL enables customers to easily adopt FPGAs and leverage the performance and power benefits the devices provide.”
Altera SDK for OpenCL Design Flow
OpenCL is an open, royalty-free standard for cross-platform, parallel programming of hardware accelerators, including CPUs, GPGPUs and FPGAs. The Altera SDK for OpenCL offers a unified, high-level design flow for hardware and software development that automates the time-consuming tasks required in typical hardware-design language (HDL) flows. The OpenCL tool flow automatically converts OpenCL kernel functions into custom FPGA hardware accelerators, adds interface IPs, builds interconnect logic and generates the FPGA programming file. The SDK includes libraries that link to OpenCL API calls within a host program running on the CPU. By automatically handling these steps, designers are able to focus their development efforts on defining and iterating their algorithms rather than designing hardware.
The portability of the OpenCL code enables users to migrate their designs to different FPGAs or SoC FPGAs as their application requirements evolve. With SoC FPGAs, the CPU host is embedded into the FPGA, providing a single-chip solution that delivers significantly higher bandwidth and lower latency between the CPU host and the FPGA compared to using two discrete devices.
Using FPGAs to Extract Maximum Parallelism in Heterogeneous Platforms
The Altera SDK for OpenCL enables programmers to leverage the massively parallel, fine-grained architectures featured in FPGAs to accelerate parallel computation. Unlike CPUs and GPGPUs, where parallel threads are executed across an array of cores, FPGAs allow kernel functions to be transformed into dedicated, deeply pipelined hardware circuits that are multithreaded using the concept of pipeline parallelism. Each of these pipelines can be replicated many times to provide even more parallelism by allowing multiple threads to execute in parallel. The result is an FPGA-based solution that can deliver >5X performance/Watt compared to alternative hardware implementations.
Altera is working with several board partners to deliver COTS board solutions to customers. Currently, boards from BittWare and Nallatech are designed to support Altera OpenCL. Additional third-party boards will be supported with future releases of the SDK.
Altera has performed a variety of benchmarks that show the productivity savings and the performance and power efficiency gained by using an OpenCL framework for FPGA development. Based on early benchmarks and working with customers in a variety of markets, the SDK shaved months off one customer’s development time for their video processing application and boosted performance by 9X versus a CPU in another customer’s financial application.
Availability
The Altera SDK for OpenCL is production ready and is available to customers through an early access program. To discover the high performance, power-efficient acceleration that OpenCL provides with FPGAs, contact a local Altera sales representative. For additional information regarding OpenCL and the benefits of targeting FPGA through an OpenCL implementation, visithttp://www.altera.com/products/software/opencl/opencl-index.html.
OpenCL for Altera FPGAs: Accelerating Performance and Design Productivity [Altera, Nov 5, 2012]
Combining the Open Computing Language (OpenCL™) programming model with Altera’s massively parallel FPGA architecture provides a powerful solution for system acceleration. The Altera® SDK for OpenCL* provides a design environment for you to easily implement OpenCL applications on FPGAs.
Benefits of OpenCL on FPGAs
Software Developer
As a software developer, how can you benefit from OpenCL on FPGAs?
As the “power wall” continues to prevent higher frequencies to be achieved in processors, multi-core processors have become the norm. This has opened the door for parallel processing techniques and thus FPGAs, which are inherently parallel, to start playing a bigger role in the embedded systems world.
The approaches to finding parallelism can be a different way of thinking for some software programmers, where FPGA designers tend to naturally think this way. You can take the scatter-gather approach for data parallelism, sending input data to the appropriate parallel resources and combining the results later, or the divide and conquer method for task parallelism, where you decompose the problem into sub problems and run them on the appropriate resources.
Using OpenCL, you continue to develop your code in the familiar C programming language but target certain functions as OpenCL kernels using the additional OpenCL constructs. Then these kernels can be sent to the available system resources, such as an FPGA, without having to learn the low level Hardware Description Language (HDL) coding practices of FPGA designers.
- HDL coding is the equivalent to coding in assembly to software developers. OpenCL keeps you in a higher level coding language that you are already familiar with, C, with some new OpenCL construct.
- Profile your code and determine the performance intensive inner loop functions that make sense to hardware accelerate as kernels in an FPGA.
- It’s about performance per watt. You’re balancing high performance with a power-efficient solution in an FPGA.
- With the FPGAs fine-grain parallelism architecture, the Altera SDK for OpenCL generates only the logic you need to deliver with as low as 1/5 of the power of other hardware alternatives.
- Kernels can target FPGAs, CPUs, GPUs, and DSPs seamlessly to produce a truly heterogamous system.
FPGA Developer
As an Embedded or DSP Designer, how can you benefit from OpenCL on FPGAs?
- Achieve significantly faster time to market compared to the traditional FPGA design flow.
- Describe your algorithms using the OpenCL C (based on ANSI C) parallel programming language instead of the traditional low-level HDL.
- Perform design exploration quickly by staying at a higher level of design abstraction.
- Obsolescence-proof your designs as you can retarget your OpenCL C code to current and future FPGAs.
- Obsolescence-proof your designs as you can retarget your OpenCL C code to current and future FPGAs.
- Generate an FPGA implementation of your OpenCL C code in a single step, bypassing the manual timing closure efforts and implementation of communication interfaces between the FPGA, host, and external memories.
The growing need for higher performance and faster time to market through parallel programming in software is seen in many markets, including the Computer & Storage, Military, Medical, and Broadcast markets.
Next Steps
- Buy a board from one of our preferred partners
- Download the Altera SDK for OpenCL
- Take an OpenCL training course
- Register for updates on Altera’s OpenCL solution for FPGAs
White Papers
- Implementing FPGA Design with the OpenCL Standard (PDF)
- Fractal Video Compression in OpenCL: An Evaluation of CPUs, GPUs, and FPGAs as Acceleration Platforms (PDF)
- Using OpenCL to Evaluate the Efficiency of CPUs, GPUs, and FPGAs for Information Filtering (PDF)
- 40Gbit AES Encryption Using OpenCL and FPGAs (PDF)
Computer and Storage [Altera, Nov 5, 2012]
Computer and storage technology is evolving rapidly. Today, cloud computing is enabling the consolidation of traditional IT functions with entirely new capabilities. For example, many large-scale data centers are now providing traditional IT services along with new data analytics services.
Hence, these large-scale data centers require highly efficient server and storage systems. Traditional CPU technology limits performance, as the use of frequency scaling as a way to increase performance has ended. The end of frequency scaling has caused a shift to multicore processing. However, multicore processing has diminishing returns in terms of increasing true application performance due to limits in I/O and memory bandwidth.
Altera® FPGAs can be used to accelerate the performance of large-scale data systems. Altera FPGAs enable higher speed data processing by providing customized high-bandwidth, low-latency connections to network and storage systems. In addition, Altera FPGAs provide compression, data filtering, and algorithmic acceleration.
With the Altera SDK for OpenCLTM, you can now rapidly develop acceleration solutions for computer and storage systems. The Altera SDK for OpenCL enables even software developers to easily design with FPGAs by allowing them to utilize a high-level programming language for developing acceleration functions.
OpenCL and the OpenCL logo are trademarks of Apple Inc. used by permission by Khronos.
OpenCL for Military [Altera, Oct 10, 2013]
Radar backend processing is a compute-intensive operation using various algorithms such as a FIR filter, which utilize custom pipeline parallelism. Increased performance is achieved by off loading from the host processor onto an FPGA.
Custom processors can be created using the OpenCL™ toolflow that are more efficient than multicore CPUs or GPUs both in computational capability and power requirements.
Figure 1: Radar Back-End Processing Alternatives Using OpenCL
For more information regarding Altera’s OpenCL for Military, please contact us at mil@altera.com.
Medical: Hardware Acceleration with OpenCL [Altera, Feb 16, 2013]
Ultrasound, X-ray, CT, and PET applications all require intensive back-end compute operations for algorithms such as fast Fourier transform (FFT) using custom pipeline parallelism. Increased algorithm performance is achieved by off loading from the host processor onto an FPGA.
Custom processors created using the OpenCL™ toolflow are more efficient than multicore CPUs or GPUs, both in computational capability and power requirements.
Related Links
OpenCL for Altera FPGAs web page
Broadcast: Advanced Systems Development Kit [Altera, Oct 25, 2012]
The Advanced Systems Development Kit is a platform that can pack multi-channel 4K video ingest, processing, and streaming into a server-ready board. It features industry-leading PCIe gen3x16 interface, plus over 1 million FPGA Logic Elements to handle the toughest video processing algorithms, matched by over 1500Gbps of external memory bandwidth – enough to tackle 4 channels of 4K UHDTV video streams. This platform provides an order of magnitude improvement in existing development kit hardware capabilities; in addition to innovations in the soft content and business model that come together to significantly accelerate end-product deployment.
Figure 1: Altera’s Advanced Systems Development Kit
Typical development kits are intended for lab-use only, because they lack the on-board resources to develop the entire end product. It is common for engineers to design their own board and software from scratch – until now. The Advanced Systems Development Kit breaks through all those barriers and significantly shortens your design cycle in many ways, including:
- A complete OmniTek BSP (board support package) for video applications, with firmware, and Windows and Linux drivers
- An evaluation design featuring OmniTek’s PCI Express DMA engine that efficiently streams multiple channels of videos between I/O and host memory
- A flexible front-panel FMC I/O expansion connector, allowing for connectivity to popular standards such as SFP+, fiber, QSFP, gigabit Ethernet, etc.
- Dual Stratix V FPGAs to integrate functions such as multi-channel format conversions, video codecs, ingest/playout connectivity, etc.
- Over 1500Gbps of external memory bandwidth – enough to handle multiple 4k channels
- PCIe gen3x16 to handle even the most demanding video streaming and acceleration
- PCIe form-factor compliant for use in both custom-built chassis and commercial off-the-shelf (COTS) servers
- Licensable full manufacturing rights to the board design, which enables you to easily make cost-optimizations and derivatives for rapid deployment of your products.
The Advanced Systems Development Kit resolves common broadcast challenges related to:
- Increased channel density
- 4K and beyond-HD resolutions
- High frame rate applications
- The fine balance between future-proofing and cost-efficiency
A rich partner ecosystem significantly accelerates and simplifies system-level advanced development. For example, Embrionix’s emSFP modules convert SDI to a number of physical layer standards, allowing you to rapidly release products and still future-proof the hardware with a simple upgrade of the emSFP. This provides a new level of flexibility for manufacturers. The combination of capabilities and physical design positions this platform perfectly for the convergence of broadcast and IT technologies.
Figure 2: Embrionix’s embedded SFP modules for high-density video connectivity
Altera’s OpenCL Toolflow
In addition to accelerating hardware designs, the Advanced Systems Development Kit will also support Altera’s unique OpenCL™ toolflow to elevate software productivity. OpenCL enables viable software implementations of complex video algorithms, and dramatically lowers the cost of the end product. Examples of broadcast applications include:
- Acquisition: Real-time debayering of raw camera data, scaling for multiviewers, etc.
- Post-production: Color grading, motion estimation, special effects rendering, etc.
- Distribution: 3D/temporal noise reduction, H.264 compression, etc.
- Consumption: JPEG2000 decoding for 4K digital cinema playout, block artifact reduction filters, etc.
The OpenCL toolflow leverages parallel processing on the underlying hardware, and achieves an order of magnitude performance improvement compared to sequential CPU processing. Furthermore, running OpenCL on the Advanced Systems Development Kit gives you several unique advantages including:
- The best performance per watt consumed, so you enjoy OpenCL’s benefits without power and heat issues from GPUs
- The ability to assimilate, manipulate, and transport multichannel video on a single board
- The highest level of integration to achieve maximum channel density for your end product
OpenCL and the OpenCL logo are trademarks of Apple Inc. used by permission by Khronos.
Related Links
EEVblog #496 – What Is An FPGA? [EEVblog YouTube channel, July 19, 2013]
5. Altera SoC FPGAs
It goes back to Altera SoC FPGAs – ARM TechCon 2011 [ARMflix YouTube channel, Oct 25, 2011]
with current state as Altera showcases Soc devices at Embedded world [embeddednewstv YouTube channel, Oct 21, 2013]
Generation 10 FPGAs and SoCs [Altera, May 16, 2013]
Altera’s Generation 10 FPGAs and SoCs optimize process technology and architecture to deliver the industry’s highest performance and highest levels of system integration at the lowest power. Initial Generation 10 families include Stratix® 10 and Arria® 10 FPGAs and SoCs with embedded processors.
Read the White paper: Expect a Breakthrough Advantage in Next-Generation FPGAs (PDF) [June 2013]
Read the White paper: Meeting the Performance and Power Imperative of the Zettabyte Era with Generation 10 (PDF) [June 2013]
Watch the video: Arria 10 FPGAs and SoCs — Reinventing the Midrange [June 2013]
Read the White paper: The Breakthrough Advantage for FPGAs with Tri-Gate Technology (PDF) [June 2013]
Generation 10 FPGAs and SoCs are supported by a leading-edge suite of development tools delivering:
- 8x improvements in compile times
- Higher level design flows that support hardware and software designers
Stratix 10 FPGAs and SoCs [Altera, June 10, 2013]
Stratix® 10 FPGAs and SoCs offer breakthrough advantages in bandwidth and system integration, including the next-generation hard processor system (HPS), to deliver the industry’s highest performance and most power- efficient FPGAs and SoCs. Stratix 10 devices are manufactured on the revolutionary Intel 14 nm 3D Tri-Gate transistor technology, which delivers breakthrough levels of performance and power efficiencies that were previously unimaginable. When coupled with 64 bit quad-core ARM® CortexTM-A53 processors and advanced heterogeneous development and debug tools such as the Altera® SDK for OpenCLTM and SoC Embedded Design Suite (EDS), Stratix 10 devices offer the industry’s most versatile heterogeneous computing platform.
White paper: The Breakthrough Advantage for FPGAs with Tri-Gate Technology [June 2013]
Industry’s First Gigahertz FPGAs and SoCs
- New ultra-high performance FPGA architecture
- 2x the core performance of prior generation high-end FPGAs
- >10 TFLOPs of single-precision floating-point DSP performance
- >4x processor data throughput of prior-generation SoCs
Break the Bandwidth Barrier with Unimaginable High-Speed Interface Rates
- 4x serial transceiver bandwidth from previous generation FPGAs for high port count designs
- 28 Gbps backplane capability for versatile data switching applications
- 56 Gbps chip-to-chip/module capability for leading edge interface standards
- Over 2.5 Tbps bandwidth for serial memory with support for Hybrid Memory Cube
- Over 1.3 Tbps bandwidth for parallel memory interfaces with support for DDR4 at 3200 Mbps
Lower Capital Expenditures (CapEx)
- Largest monolithic FPGA device with >4M logic elements offer an unprecedented level of integration capability
- Heterogeneous multi-die 3D solutions including SRAM, DRAM, and ASICs
- Next-generation HPS
Lower Operating Expenses (OpEX)
- Leveraging Intel’s leadership in process technology, Stratix 10 FPGAs offer the most power-efficient technologies
- 70% lower power than prior generation high-end FPGAs and SoCs
- 100 GFlops/Watt of single-pecision floating point efficiency
- Integrated host processor for operation, administration, and maintenance minimizes system down time
Versatile Heterogeneous Computing for Performance and Power-Efficient SoC Design
- 64 bit quad-core ARM Cortex-A53 processor optimized for ultra-high performance per watt
- Heterogeneous C-based modeling and hardware design with Altera SDK for OpenCL
- Heterogeneous debug, profiling, and whole chip visualization with Altera SoC EDS featuring ARM Development Suite™ (DS-5™) Altera Edition Toolkit
Reduce Time-to-Market
- Fastest compile times in the industry
- C-based design entry using the Altera SDK for OpenCL, offering a design environment that is easy to implement on FPGAs
- Start developing with Arria 10 devices and then migrate to footprint-compatible Stratix 10 devices
- Complementary Enpirion PowerSoCs will offer customers higher performance, lower system power, higher reliability, smaller footprint, and faster time-to-market to power Stratix 10 FPGAs and SoCs
Altera to Build Next-Generation, High-Performance FPGAs on Intel’s 14nm Tri-Gate Technology
Table 1. Stratix 10 Family Variants
Stratix 10 FPGAs and SoC family is ideal to meet your high-performance, high-bandwidth, and low power requirements in the communication infrastructure, cloud computing and data centers, high-performance computing, military, broadcast, test and measurement, and other applications.
Related Links
- White paper: The Breakthrough Advantage for FPGAs with Tri-Gate Technology (PDF) [June 2013]
- White paper: Expect a Breakthrough Advantage in Next Generation FPGAs (PDF) [June 2013]
- White paper: Meeting the Power and Performance Imperative of the Zettabyte Era with Generation 10 (PDF) [June 2013]
- Press Release: Altera Announces Breakthrough Advantage with Generation 10 FPGAs and SoCs [June 2013]
- Generation 10 Portfolio
Arria 10 SoC [Altera, June 10, 2013]
Arria 10 SoCs: Reinventing the Midrange
The 20 nm Arria® 10 ARM-based SoCs deliver optimal performance, power efficiency, small form factor, and low cost for midrange applications. Arria 10 SoCs, based on TSMC’s 20 nm process technology, combine a dual-core ARM® Cortex™-A9 MPCore™ hard processor system (HPS) with industry-leading programmable logic technology. Arria 10 SoCs offer a processor with a rich feature set of embedded peripherals, variable-precision digital signal processing (DSP) blocks, embedded high-speed transceivers, hard memory controllers, and protocol IP controllers – all in a single highly integrated package.
Arria 10 SoCs: Across-the Board Improvements
Arria 10 SoCs combine architectural innovations with TSMC’s 20 nm process technology to deliver improvements in performance and power reduction:
- 87% higher processor performance with up to 1.5 GHz CPU operation per core
- 60% higher performance versus the previous generation, over 500 MHz-capable core performance (15% higher performance than previous SoC)
- 4X more transceiver bandwidth versus the previous generation (2X more bandwidth versus previous high-end FPGAs)
- 4X higher system performance (2666 Mbps DDR4, Hybrid Memory Cube support)
- More than 3300 18×19 multipliers implemented on variable-precision DSP
- 40% lower power with process technology improvement and innovative techniques for power reduction
Table 1. Arria SoC Feature Comparison
Note: See full list of memory devices supported
Designed for Productivity
Design productivity is one of the driving philosophies of the Arria 10 SoC architecture. Arria 10 SoC offer full software compatibility with previous generation SoCs, a broad ecosystem of ARM software and tools, and the enhanced FPGA and DSP hardware design flow.
- Extensive ecosystem of ARM for software development
- Altera SoC Embedded Design Suite featuring the ARM Development Studio 5 (DS-5™) Altera Edition Toolkit
- Board support packages for popular operating system including Linux, Wind River’s VxWorks, Micro-C OS II, and more
- Full software compatibility between 28 nm Cyclone V and Arria V SoCs and Arria 10 SoCs
- Quartus® II FPGA Design Suite featuring:
- High-level automated design flow with OpenCL™ compiler from Altera
- Model-based DSP hardware design with Altera DSP Builder
Target Markets
Arria 10 SoCs have been designed to meet the performance, power, and cost requirements for applications such as:
- Wireless infrastructure equipment including remote radio unit and mobile backhaul
- Compute and storage equipment including flash cache, cloud computing, and acceleration
- Broadcast studio and distribution equipment including professional A/V and video conferencing
- Military guidance, control and intelligence equipment
- Wireline 100G line cards, bridges and aggregation, 40G GPON
- Test and measurement equipment
- Diagnostic medical imaging equipment
Related Links
- Arria 10 Advance Information Brief (PDF)
- White paper: Meeting the Performance and Power Imperative of the Zettabyte Era with Generation 10 (PDF) [June 2013]
- White paper: Expect Breakthrough Capabilities in Next Generation FPGAs (PDF) [June 2013]
- Video: Arria 10 FPGAs and SoCs – Reinventing the Midrange [June 2013]
- SoC overview [June 2013]
With Android and forked Android smartphones as the industry standard Nokia relegated to a niche market status while Apple should radically alter its previous premium strategy for long term
Here is the chart reflecting the performance of the market-leading mobile phones upto Q2’13:
From this the most visible things are:
- Android and Android-forked (Xiaomi etc.) smartphones are the undisputed industry standards to dominate the market in years to come
- Both the Symbian to Windows Phone and S40 to Asha Full Touch smartphone platform transition strategies from Nokia could survive the continued Android onslaught but only in a niche market status
- There is no room for Apple’s further growth, and both the platform and the company could face a gradual decline in the smartphone market
My other observations about the state of the smartphone market after Q2’13 were already presented in the following posts:
- Superphones turning point: segment satured with Tier 1 globals while the Chinese locals are at less than 40% of the Samsung price [‘Experiencing the Cloud’, Aug 3, 2013] OR Samsung is leapfrogging Apple while the Chinese local brands are coming close to Samsung but at less than 40% price. Meanwhile the superphone segment of the market becomes saturated.
- Xiaomi, OPPO and Meizu–top Chinese brands of smartphone innovation [‘Experiencing the Cloud’, Aug 1, 2013]
- GiONEE (金立), the emerging global competitor on the smartphone market [‘Experiencing the Cloud’, July 22, 2013]
- Eight-core MT6592 for superphones and big.LITTLE MT8135 for tablets implemented in 28nm HKMG are coming from MediaTek to further disrupt the operations of Qualcomm and Samsung [‘Experiencing the Cloud’, July 20-29, 2013]
- China: Entry-level dual core IPS WVGA (480×800) smartphones $65+ now, quad-core $70+ in June [‘Experiencing the Cloud’, April 29, 2013]
In essence we came to a point when the superphone market came down in price to as low as $110 and up, while the entry-level segment of good quality came down to a $65+ price level. Also the smartphone market became saturated in all segments which brings an end to Samsung’s ability to base its premium profitability ambitions on smartphones alone (almost), as it was reflected in 20 years of Samsung “New Management” as manifested by the latest, June 20th GALAXY & ATIV innovations [‘Experiencing the Cloud’, July 2-26, 2013]:
… innovations in the broadest sense of the world: technology, hardware and software engineering and design, marketing in general and branding in particular etc.
Updates: Q2 record-high operating profit + smartphone worries deepen + overall business situation + nonproportionally high capex of the semiconductor business + the #2 capex beneficiary, the Display Panel Segment
These observations also led to much greater conclusions about the upcoming changes:
- China is the epicenter of the mobile Internet world, so of the next-gen HTML5 web [‘Experiencing the Cloud’, Aug 5, 2013]
- The Upcoming Mobile Internet Superpower [‘Experiencing the Cloud’, Aug 13, 2013]
Below I will assess the ‘Nokia Q2’13 market situation and changes’ as well as include ‘Gartner’s own assessment of the Q2’13 overall market situation and the changes’ to complete the picture.
Nokia Q2’13 market situation and changes:
Looking at the progress of Nokia Symbian to Windows Phone transformation Q2’13 was a straight continuation of the trends noted for Q1’13 in Nokia: Continued moderate progress with Lumia, urgent Asha Touch refresh and new innovations to come against the onslaught of unbranded Android and forked Android players in China and India [‘Experiencing the Cloud’, April 18, 2013] as you could also well observe from the chart included here as well:
Nokia was extensively discussing its Windows Phone transition in Nokia Corporation Interim Report for Q2 2013 and January-June 2013 [press release, July 18, 2013]:
-
Lumia Q2 volumes increased 32% quarter-on-quarter to 7.4 million units, reflecting strong demand from customers for a broadened Lumia product range.
-
Commenting on the second quarter results, Stephen Elop, Nokia CEO, said: “ … In our Smart Devices business unit, we continue to focus on delivering meaningful differentiation to consumers around the world. We are very proud of the recent creations by our Lumia team, from the Lumia 520 – our most affordable Windows Phone 8 product which has enjoyed a strong start in markets like China, France, India, Thailand, the UK, the US and Vietnam – to the Lumia 1020, our star imaging product which we unveiled to the world last week. Overall, Lumia volumes grew to 7.4 million in the second quarter, the highest for any quarter so far and showing increasing momentum for the ecosystem. During the third quarter, we expect that our new Lumia products will drive a significant part of our Smart Devices revenue.”
-
In the third quarter 2013, supported by the wider availability of recently announced Lumia products as well as recently announced Mobile Phones products, Nokia expects higher Devices & Services net sales, compared to the second quarter 2013.
-
The year-on-year decline in our Smart Devices volumes in the second quarter 2013 continued to be driven by the strong momentum of competing smartphone platforms and our portfolio transition from Symbian products to Lumia products. The decline was primarily due to lower Symbian volumes, partially offset by higher Lumia volumes. Our Symbian volumes decreased from 6 million units in the second quarter 2012 to approximately zero in the second quarter 2013. Our Lumia volumes increased from 4.0 million in the second quarter 2012 to 7.4 million in the second quarter 2013.
-
On a sequential basis, the increase in our Smart Devices volumes in the second quarter 2013 was due to higher Lumia volumes, as we started shipping the Lumia 520 and 720 in significant volumes. In the second quarter 2013, the vast majority of Smart Devices volumes were from Windows Phone 8-based Lumia products.
-
The year-on-year increase in our Smart Devices ASP in the second quarter 2013 was primarily due to a positive mix shift towards sales of our Lumia products which carry a higher ASP than our Symbian products, partially offset by our pricing actions. Sequentially, the decrease in our Smart Devices ASP in the second quarter 2013 was primarily due to a negative mix shift towards sales of our lower priced Windows Phone 8-based Lumia products as well as our pricing actions.
-
Nokia announced and started shipments in select markets of the Nokia Lumia 925, a new interpretation of its award-winning flagship, the Nokia Lumia 920. The Nokia Lumia 925 introduces metal for the first time to the Nokia Lumia range and includes the most advanced lens technology and next-generation imaging software to capture clearer and sharper pictures and video even in low light conditions. The Nokia Lumia 925 offers a variety of exclusive services such as Nokia Music for unlimited streaming of free playlists, integrated HERE services, and the option to add wireless charging with a snap-on wireless charging cover.
-
Nokia announced the Nokia Lumia 928 smartphone, exclusive to Verizon Wireless. With a 8.7MP camera and Nokia’s PureView imaging innovation, the Nokia Lumia 928 delivers superior imaging and video performance that enables people to capture bright, blur free photos and videos, even in low light conditions. The sleek and stylish smartphone comes with the latest high-end Nokia Lumia experiences, including Nokia Music, HERE services, and built-in wireless charging.
-
Nokia started shipping in volumes the Nokia Lumia 520, its most affordable Windows Phone 8 smartphone, delivering experiences normally found only in high-end smartphones, such as the same digital camera lenses found on the Nokia Lumia 920, Nokia Music for free music out of the box and even offline, and HERE services.
-
Nokia’s Lumia range of smartphones continued to attract businesses, including Miele & Cie. KG, a global leader in domestic appliances and commercial machinery, which has chosen the Nokia Lumia range as the smartphone of choice for its global employees.
-
The Windows Phone Store continued to strengthen in terms of the quantity and quality of applications. The Windows Phone Store today offers more than 165 000 applications and games.
The Q2’13-related improvements mentioned above and influencing the below chart were even more extensively discussed in my earlier posts:
-
High-volume Nokia Lumia superphones with Windows Phone 8 extended on the top for China, and on the entry level needed for Asia and Middle-East as well UPDATE: at even lower price by 27% [‘Experiencing the Cloud’, Dec 5, 2012 – March 21, 2013] Note that the Lumia 520 W-CDMA mentioned there for ¥ 1299.00 [$209] is now (Aug 17) ¥ 899.00 [$147] while in India it is even lower priced at Rs 7,667+ [$124+]
-
Nokia’s expanded, new risks and uncertainties for its Windows Phone strategy for 2013 [‘Experiencing the Cloud’, March 17, 2013]
while the Q3’13-related actions of improvements in these posts:
-
Nokia Lumia 1020: an excellent case of Nokia’s contribution to Microsoft as a key innovation partner [‘Experiencing the Cloud’, July 12, 2013]
-
Minutes of a high-octane but also expert evangelist CEO: Stephen Elop, Nokia [‘Experiencing the Cloud’, July 12, 2013]
Now look again at the performance chart for the reflections:
From the further decline of Asha Full Touch you could see that the Temporary Nokia setback in India [‘Experiencing the Cloud’, April 28, 2013] continued into the Q2’13 as well as the result of entry-level local brand Android smartphones being in heavy price competition with Nokia Asha Full Touch during Q2 while having superior hardware specifications. Even Samsung’s REX 70 competed in price with Asha Full Touch.
Nokia was talking in his Nokia Corporation Interim Report for Q2 2013 and January-June 2013 [press release, July 18, 2013] only about the following future-oriented actions that were introduced in Q2 in order to remedy this situation:
In Devices & Services, our Mobile Phones business unit started to demonstrate some signs of recovery in the latter part of the second quarter following a difficult start to the year. Also, towards the end of the second quarter, we started to ship the Asha 501, which brings a new design and user experience to the highly competitive sub-100 USD market. While we are very encouraged by the consumer response to our innovations in this price category, our Mobile Phones business unit is planning to take actions to focus its product offering and improve product competitiveness.
On a year-on-year basis, our Mobile Phones volumes in the second quarter 2013 were negatively affected by competitive industry dynamics, including intense smartphone competition at increasingly lower price points and intense competition at the low end of our product portfolio. Compared to the second quarter 2012, our Mobile Phones volumes declined across our portfolio, most notably for our non-full-touch devices that we sell to our customers for above EUR 30, partially offset by higher sales volumes of Asha full-touch smartphones.
Nokia started production at its new manufacturing facility in Hanoi, Vietnam. The new site has been established to produce our most affordable Asha smartphones and feature phones.
Nokia announced and started shipments of the Nokia Asha 501, the first of a new generation of smartphones to run on the new Asha platform. Retailing at a suggested price of USD 99, the Nokia Asha 501 offers users affordable smartphone design with bold color, a high-quality build and an innovative user interface. The new Asha platform also allows developers who write applications for the Nokia Asha 501 to reach all smartphones based on the new Asha platform without having to re-write code.
These things were already extensively discussed in my earlier posts:
- Nokia’s non-Windows crossroad [‘Experiencing the Cloud’, May 2, 2013]
- New Asha platform and ecosystem to deliver a breakthrough category of affordable smartphone from Nokia [‘Experiencing the Cloud’, May 9, 2013] my composite post of the all relevant launch information
- New Nokia Asha platform for developers [‘Experiencing the Cloud’, May 9, 2013] my composite post of the all relevant development platform information
- Nokia becoming the next Samsung from its new Vietnamese manufacturing base? [‘Experiencing the Cloud’, June 24, 2013]
And here is how Gartner was assessing the Q2’13 overall market situation and the changes:
Gartner Says Smartphone Sales Grew 46.5 Percent in Second Quarter of 2013 and Exceeded Feature Phone Sales for First Time [press release, Aug 14, 2013]
- Worldwide Mobile Phone Sales Grew 3.6 Percent in Second Quarter of 2013
- Microsoft Has Become the No. 3 Smartphone OS Overtaking BlackBerry
Worldwide mobile phone sales to end users totaled 435 million units in the second quarter of 2013, an increase of 3.6 percent from the same period last year, according to Gartner, Inc. Worldwide smartphone sales to end users reached 225 million units, up 46.5 percent from the second quarter of 2012. Sales of feature phones to end users totaled 210 million units and declined 21 percent year-over-year.
“Smartphones accounted for 51.8 percent of mobile phone sales in the second quarter of 2013, resulting in smartphone sales surpassing feature phone sales for the first time,” said Anshul Gupta, principal research analyst at Gartner. Asia/Pacific, Latin America and Eastern Europe exhibited the highest smartphone growth rates of 74.1 percent, 55.7 percent and 31.6 percent respectively, as smartphone sales grew in all regions.
Samsung maintained the No. 1 position in the global smartphone market, as its share of smartphone sales reached 31.7 percent, up from 29.7 percent in the second quarter of 2012 (see Table 1). Apple’s smartphone sales reached 32 million units in the second quarter of 2013, up 10.2 percent from a year ago.
Table 1
Worldwide Smartphone Sales to End Users by Vendor in 2Q13 (Thousands of Units)
Company
2Q13 Units
2Q13 Market Share (%)
2Q12 Units
2Q12 Market Share (%)
Samsung
71,380.9
31.7
45,603.8
29.7
Apple
31,899.7
14.2
28,935.0
18.8
LG Electronics
11,473.0
5.1
5,827.8
3.8
Lenovo
10,671.4
4.7
4,370.9
2.8
ZTE
9,687.6
4.3
6,331.4
4.1
Others
90,213.6
40.0
62,704.0
40.8
Total
225,326.2
100.0
153,772.9
100.0
Source: Gartner (August 2013)
In the smartphone operating system (OS) market (see Table 2), Microsoft took over BlackBerry for the first time, taking the No. 3 spot with 3.3 percent market share in the second quarter of 2013. “While Microsoft has managed to increase share and volume in the quarter, Microsoft should continue to focus on growing interest from app developers to help grow its appeal among users,” said Mr. Gupta. Android continued to increase its lead, garnering 79 percent of the market in the second quarter.
Table 2
Worldwide Smartphone Sales to End Users by Operating System in 2Q13 (Thousands of Units)
Operating System
2Q13 Units
2Q13 Market Share (%)
2Q12 Units
2Q12 Market Share (%)
Android
177,898.2
79.0
98,664.0
64.2
iOS
31,899.7
14.2
28,935.0
18.8
Microsoft
7,407.6
3.3
4,039.1
2.6
BlackBerry
6,180.0
2.7
7,991.2
5.2
Bada
838.2
0.4
4,208.8
2.7
Symbian
630.8
0.3
9,071.5
5.9
Others
471.7
0.2
863.3
0.6
Total
225,326.2
100.0
153,772.9
100.0
Source: Gartner (August 2013)
Mobile Phone Vendor Perspective
Samsung: Samsung remained in the No. 1 position in the overall mobile phone market, with sales to end users growing 19 percent in the second quarter of 2013 (see Table 3). “We see demand in the premium smartphone market come mainly from the lower end of this segment in the $400-and-below ASP mark. It will be critical for Samsung to step up its game in the mid-tier and also be more aggressive in emerging markets. Innovation cannot be limited to the high end,” said Mr. Gupta.
Nokia: Slowing demand of feature phone sales across many markets worldwide, and fierce competition in the smartphone segment, affected Nokia’s mobile phone sales in the second quarter of 2013. Nokia’s mobile phone sales totaled 61 million units, down from 83 million units a year ago. Nokia’s Lumia sales grew 112.7 percent in the second quarter of 2013 thanks to its expanded Lumia portfolio, which now include Lumia 520 and Lumia 720. “With the recent announcement of the Lumia 1020, Nokia has built a wide portfolio of devices at multiple price points, which should boost Lumia sales in the second half of 2013,” said Mr. Gupta. “However, Nokia is facing tough competition from Android devices, especially from regional and Chinese manufacturers which are more aggressive in terms of price points.”
Apple: While sales continued to grow, the company faced a significant drop in the ASP of its smartphones. Despite the iPhone 5 being the most popular model, its ASP declined to the lowest figure registered by Apple since the iPhone’s launch in 2007. The ASP reduction is due to strong sales of the iPhone 4, which is sold at a strongly discounted price. “While Apple’s ASP demonstrates the need for a new flagship model, it is risky for Apple to introduce a new lower-priced model too,” said Mr. Gupta. “Although the possible new lower-priced device may be priced similarly to the iPhone 4 at $300 to $400, the potential for cannibalization will be much greater than what is seen today with the iPhone 4. Despite being seen as the less expensive sibling of the flagship product, it would represent a new device with the hype of the marketing associated with it.”
Lenovo: Lenovo’s mobile phone sales grew 60.6 percent to reach 11 million units in the second quarter of 2013. Lenovo’s quarter performance was bolstered by smartphone sales. Its smartphone sales grew 144 percent year-over-year and helped it rise to the No. 4 spot in the worldwide smartphone market for the first time. Lenovo continues to rely heavily on its home market in China, which represents more than 95 percent of its sales. It remains challenging for Lenovo to expand outside China as it has to strengthen its direct channel as well as its relationships with communications service providers.
Table 3
Worldwide Mobile Phone Sales to End Users by Vendor in 2Q13 (Thousands of Units)
Company
2Q13 Units
2Q13 Market Share (%)
2Q12 Units
2Q12 Market Share (%)
Samsung
107,526.0
24.7
90,432.1
21.5
Nokia
60,953.7
14.0
83,420.1
19.9
Apple
31,899.7
7.3
28,935.0
6.9
LG Electronics
17,016.4
3.9
14,345.4
3.4
ZTE
15,280.7
3.5
17,198.2
4.1
Huawei
11,275.1
2.6
10,894.2
2.6
Lenovo
10,954.8
2.5
6,821.7
1.6
TCL Communi-cation [Alcatel]
10,134.3
2.3
9,355.7
2.2
Sony Mobile Communications
9,504.7
2.2
7,346.8
1.7
Yulong [Coolpad]
7,911.5
1.8
4,016.2
1.0
Others
152,701.5
35.1
147,354.60
35.1
Total
435,158.4
100.0
420,120.0
100.0
Source: Gartner (August 2013)
“With second quarter of 2013 sales broadly on track, we see little need to adjust our expectations for worldwide mobile phone sales forecast to total 1.82 billion units this year. Flagship devices brought to market in time for the holidays, and the continued price reduction of smartphones will drive consumer adoption in the second half of the year,” said Mr. Gupta.
Additional information is in the Gartner report “Market Share Analysis: Mobile Phones, Worldwide, 2Q13.” The report is available on Gartner’s website at http://www.gartner.com/document/2573119.