All posts by admin

Cascading 2.2 Now Available

Introducing Cascading 2.2, the leading Java application framework for building and deploying reliable enterprise Big Data applications on Hadoop

Introducing Cascading 2.2

Cascading is a Java application framework that enables typical developers to quickly and easily develop rich Data Analytics and Data Management applications that can be deployed and managed across a variety of computing environments. Cascading works seamlessly with Apache Hadoop 1.0 and API compatible distributions.

Cascading 2.2 is now publicly available for download.

http://www.cascading.org/downloads/

Cascading 2.2 includes a number of new features and updates:

What’s New in Cascading 2.2
– First class support for field level type information used by the planner
– Pluggable API for custom type coercion on custom types
– Support for blocks of Java “scripts” in addition to expressions
– AssemblyPlanner interface to allow for platform independent generative Flow planning
– Optional CombinedInputFormat support to improve handling with lots of small files
– Added FirstNBuffer and updated Unique to leverage it for faster performance
– MaxValue and MinValue Aggregators to allow for max/min on Comparable types, not just numbers

What’s Improved in Cascading 2.2
– Updated relevant operations to optionally honor SQL like semantics with regard to NULL values
– Updated SubAssembly to support setting local and step properties via the ConfigDef interface
– Updated FlowDef to accept classpath elements for dynamic inclusion of Hadoop jobs
– Updated GlobHfs to minimize resource and CPU usage

For more details on new features and resolved issues see:
https://github.com/Cascading/cascading/blob/2.2/CHANGES.txt

Supporting Resources

Availability and Pricing

Cascading 2.2 is available now, and freely licensable under the Apache 2.0 License Agreement. Concurrent offers standard and premium support subscriptions for enterprise use, with pricing based on number of users. To learn more about Concurrent’s offerings please visit http://www.concurrentinc.com/newsletter.

Open-Source Systems You May Have Taken for Granted: 10 Examples

Open-Source Systems You May Have Taken for Granted: 10 Examples
By Chris Preimesberger
October 4, 2013
http://www.eweek.com/developer/slideshows/open-source-systems-you-may-have-taken-for-granted-10-examples.html/

A key moment in IT history took place in Mountain View, Calif., on Feb. 3, 1998. That was the day a small group of Silicon Valley software developers (which included Dr. Larry Augustin, now CEO of SugarDB, Eric Raymond and Christine Peterson) sat down to decide that there needed to be an actual name for a new software development genre. The now-familiar term “open source” was first coined at this meeting. Since that day, the open-source movement—led today by Linus Torvalds’ Linux—has evolved to affect every corner of the IT industry. Open source now is more mature than ever, with today’s landscape encompassing a much broader range of applications benefiting industry and society. From Web servers to big data analytics to cloud computing to research and exploration, the following is a look at some of the most innovative open-source technologies and projects that illustrate just how far the movement has come. In this slide show, eWEEK and Concurrent CEO Gary Nakamura provide a listing of 10 well-known IT use cases that rely heavily—or even solely—on open-source software.

See the slideshow at: http://www.eweek.com/developer/slideshows/open-source-systems-you-may-have-taken-for-granted-10-examples.html/

Concurrent, Inc. and Think Big Analytics Partner to Unleash the Power of Big Data in the Enterprise

Collaboration Supports Ever-Growing Cascading Community; Offers Services to Drive Further Business Value from Big Data Application Developments on Apache Hadoop

SAN FRANCISCO – Oct. 2, 2013 – Concurrent, Inc., the enterprise Big Data application platform company, today announced a new partnership with Think Big Analytics, a provider of data science and engineering services for Big Data and big analytics projects. Now, enterprises can leverage the power of the Cascading application framework and have access to the Think Big team of expert Big Data engineers, architects and data scientists to drive further business value from their Big Data deployments on Apache Hadoop™.

Enterprises are moving from Big Data as a project to being in the “business of data,” and are leveraging Cascading to process and analyze exponentially growing volumes of data. Big Data has broad reaching and significant untapped value for the enterprise.

While Cascading allows users to leverage existing skills and IT investments to instantly create and run Big Data applications on Hadoop, enterprises often require external support to integrate Cascading into their infrastructure. As a result, Concurrent partnered with Think Big, which is known for its strategic planning, implementation and training expertise. Think Big provides valuable expertise to execute a Big Data strategy quickly and with low risk. The Big Data consulting firm provides services that ensure enterprises have the right foundation in place to make the most of their Big Data strategies. Think Big will provide customized consulting and implementation services to support enterprises’ Big Data application deployments on Hadoop in order to maximize each project’s business value.

Supporting Quotes

“In order to achieve Big Data breakthroughs, companies need to invest in both technology and people. Our partnership with Concurrent addresses both aspects as enterprises can tap into the power of the Cascading application framework while accessing the expertise and training of the Think Big team. After looking at other approaches, we’ve decided to support Cascading as it offers a proven solution for Big Data application development on Hadoop by solving real business problems at enterprise scale.”

-Ron Bodkin, Founder and CEO, Think Big Analytics

“Cascading is the most widely used and deployed application framework for building robust, enterprise Big Data applications. And, at Concurrent, we are committed to offering the best resources available to our ever-expanding user base. Our partnership with Think Big is a reflection of that commitment. Think Big offers invaluable support to our enterprise users and shares our mission of driving business differentiation through data.”

-Gary Nakamura, CEO of Concurrent, Inc.

Supporting Resources

About Think Big Analytics
Think Big Analytics is purpose-built to enable customers to create value from Big Data. The company provides data science and engineering services to assemble custom solutions that deliver business outcomes. Think Big collaborates with leaders to prioritize initiatives and generate value quickly with its proven Cornerstone Methodology (SM). The company was founded by former Quantcast and C-Bridge executives and is based in Mountain View, CA. For more information, visit www.thinkbiganalytics.com.

About Concurrent, Inc.

Concurrent, Inc. is the enterprise Big Data application platform company. Founded in 2008, Concurrent simplifies Big Data application development, deployment and management on Apache Hadoop. We are the company behind Cascading, the most widely used and deployed technology for building Big Data applications with more than 90,000 user downloads a month. Enterprises including Twitter, eBay, The Climate Corporation and Etsy all rely on Concurrent’s technology to drive their Big Data deployments. Concurrent is headquartered in San Francisco. Visit Concurrent online at http://concurrentinc.com.

 

Media Contact
Danielle Salvato-Earl
Kulesa Faul for Concurrent, Inc.
(650) 340 1982
concurrent@kulesafaul.com

How to Escape the Dark Valley of Your Hadoop Journey

How to Escape the Dark Valley of Your Hadoop Journey
Gary Nakamura, CEO, Concurrent, Inc.
September 30, 2013
http://allthingsd.com/20130930/how-to-escape-the-dark-valley-of-your-hadoop-journey/

It happens to the best of us. You know your business is bursting with useful data, and you’ve only begun to scratch the surface. So you strike out to build an analytical platform, using all the great open-source tools you’ve been hearing so much about. First, you have to capture all the data that’s coming in, before you even know what you’re going to do with it. So you build a butterfly net to catch it all, using Hadoop. But as soon as the net is cast, everything goes dark. You know the data’s there, but you can’t get at it, or if you can, it comes out in unusable formats. Your current systems won’t talk to it, and you don’t have a staff of PhDs in programming, the budget to buy a United Nations’ worth of translators or hire an army of consultants. A chill runs down the back of your neck. What have you done? You’ve entered the Dark Valley of Hadoop. That’s the bad news. The good news is that you’re not alone, and there’s a way out.

Warning Signs That You’re in the Dark Valley

Many data-rich companies fall into the Dark Valley for a time. You have the data, but you’re not getting the value that you expect from it. You have problems testing and deploying the applications that are supposed to extract that value. You have difficulty turning business requirements into code that will turn the great Leviathan that is the Hadoop Distributed File System into something halfway manageable. The project for which all of this effort was intended is delayed for months, with cost overruns making stakeholders nervous. Once you finally get the chance to test, you don’t get the results you were expecting. More delays ensue.

One of the cruelest tricks of the Dark Valley is the illusion that you got it right on the first try. Agile design philosophy tells us to work on small projects and test them as quickly as possible, then iterate forward. But Hadoop tends to reveal its weaknesses around manageability the deeper you get into the adoption cycle. If you’re using tools made for programmers, such as Pig and Hive, you’re making a bet that the programmer who built that first iteration is going to be around for the second. In today’s competitive marketplace, there is no guarantee of that. Then there is the fact that MapReduce, Hadoop’s native language, is already on its second version, with a third entirely new computation engine, built from the ground up, rapidly on its way. In the Hadoop ecosystem, many low-level moving parts have the nasty habit of changing every 90 to 120 days. All these moving parts mean that you’re having to keep up with numerous release cycles, which takes your focus off the business at hand.

So if the project mantra is “stand up, rinse and repeat,” it’s the “repeat” part that proves challenging. You find you need to expand your team, the number of systems and the scope of your project. The farther along the path you travel from “build” to “deploy and operate,” the wider the gap between programming tools and enterprise application tools becomes. The open source tools for Hadoop were simply never intended for the mainstream data-driven enterprise. The result is a skills bottleneck, requirements confusion and maintenance complexity — a lot of bumping around in the dark.

You aren’t alone. Some of the largest and most successful data-driven companies are having or have had similar frustrations. LinkedIn, Twitter and BlueKai all started out with native interfaces and have ended up with mountains of unmaintainable code. They have found better, more useable, more sustainable technologies to run their businesses. By investing time and money in alternatives that increased the productivity of their brainy staff, they fought their way out with significant investments of time and money. The good news is that you can learn from their experience and avoid the Dark Valley entirely.

Escape From the Dark Valley

There is a light at the end of the tunnel, but you have to know how to find it. The key lies in knowing your options, which usually involves leveraging the skillsets you already have in-house so that your big data strategy can continue up into the light.

The development methodologies you have in place were established for a reason. Because Hadoop wasn’t designed for the enterprise developer, the first mistake many enterprises make is to invert their planning processes around Hadoop. The best way to protect your existing methodologies is to avoid reconfiguring them in order to use MapReduce. In other words, is it possible to find an easier way for your experienced enterprise developers to use Hadoop instead of hiring MapReduce programmers and attempting to teach them about your business?

Indeed, Hadoop can be tamed through application programming interfaces (APIs) or domain-specific languages (DSLs) that encapsulate and hide MapReduce so that your developers don’t have to master it. For example, a modeling tool such as MicroStrategy, SAS, R, or SPSS can use a DSL that will consume models and run them on Hadoop without needing to write Hive, Pig or MapReduce. Enterprises can leverage existing investments made in Java, SQL and modeling tools that will allow them to quickly start parsing Hadoop datasets without the need to learn another language.

Here are some domain-specific languages that leverage existing development skills:

  • Java: Cascading is a widely used Java development framework that enables developers to quickly build workflows on Hadoop.
  • SQL: Cascading Lingual is an ANSI SQL DSL. This means that developers can leverage existing ETL and SQL and migrate them to Hadoop. The deeper benefit is that applications designed for SQL can access data on Hadoop without modification using a standard database driver.
  • MicroStrategy, SAS, R and SPSS: Cascading Pattern is a scoring engine for modeling applications. In a matter of hours, developers can test and score predictive models on Hadoop.

Newer languages like Scala and Clojure are popular with many developers these days (Clojure is especially popular with data scientists). DSLs for these languages also simplify development on Hadoop.

These APIs abstract the complexity of Hadoop, enabling enterprise developers to spend their time on business logic and creating big data applications instead of getting stuck in the maze of the Hadoop open source projects. The best APIs let you corral the resources you already know how to use — relational databases, ERP systems, and visualization tools — and use them in conjunction with Hadoop.

Conclusion

The power of big data has been established, but our understanding of how to exploit it in the most productive way is still maturing. The initial toolset that came with Hadoop didn’t anticipate the kinds of enterprise applications and powerful analyses that businesses would want to build on it. Thus, many have fallen into the Dark Valley. But a new breed of middleware (APIs and DSLs) has arrived. They keep track of all the variables and peculiarities of Hadoop, abstract them away from development, and offer better reliability, sustainability and operational characteristics so that enterprises can find their way back out into the light.

Nearly 20 years ago, the Web set the stage for existing enterprises and new emerging companies to change and create new innovative businesses. Similarly, big data and the business opportunity that it offers is driving enterprises to extract valuable insights about their business and in many cases create additional and significant monetization opportunities for their existing products and services. Enterprises are transitioning from big data as a project to being in the “business of data.” If that’s not a bright light at the end of the tunnel, what is?

Concurrent, Inc. Fulfills Mission to Bring Rapid Application Development and Deployment on Hadoop to the Masses

    • Cascading production base grows to more than 4,500 deployments, 30,000 users and 90,000 downloads per month
  • Continued innovation and strategic partnerships helps propel momentum
  • New CEO, Vice President of Engineering appointments and Series A funding support growth and product development

SAN FRANCISCO – Aug. 13, 2013 – Concurrent, Inc., the enterprise Big Data application platform company, wraps up the first half of 2013 with unprecedented growth and momentum. The company continues to drive Big Data application development on Apache Hadoop™ with the rapid adoption of its Cascading™application framework, the introduction of open source projectsCascading Lingual and Cascading Pattern, the expansion of the company’s leadership team, new strategic partnerships and $4 million in Series A funding.

Cascading is rapidly becoming the framework of choice for tackling enterprise Big Data processing needs in an effort to make the most of growing volumes of unstructured and semi-structured data stored in Hadoop. With the continuing and rapid evolution of Hadoop and the business imperative to derive more from data, Cascading provides a practical and useful abstraction for enterprises to leverage existing skill sets and easily integrate Hadoop with existing data management and analytics systems. Concurrent continues to deliver on its promise to make development and deployment of ETL, data processing and analytic applications on Hadoop easy.

“Enterprises continue to rely on Cascading as critical infrastructure to streamline and operationalize their Big Data strategy,” said Gary Nakamura, CEO of Concurrent. “With application development front and center in the world of Big Data, Cascading delivers a complete toolbox for enterprises to make the most of the business of data. We look forward to Cascading’s continued momentum in the market.”

Corporate, Project and Partnership Milestones

● Secured $4 million in Series A funding in March 2013, following a $900,000 seed investment in August 2011.
● Expanded leadership team with the appointments of Gary Nakamura as CEO, and Paul O’Leary as vice president of engineering.
● Launched Cascading 2.1, the most widely used and deployed application framework for building robust, enterprise Big Data applications on Hadoop.
● Launched Cascading Lingual, a free open source project that delivers ANSI-SQL technology to easily build new and integrate existing applications onto Hadoop.
● Launched Cascading Pattern, a free, open source, standard-based scoring engine that enables analysts and data scientists to quickly deploy machine-learning applications via PMML or an API.
● User downloads of Cascading have surpassed more than 90,000 per month.
● Addition of strategic partnerships with technology leadersHortonworks and MapR.

Supporting Resources

● Enterprise Data Workflows with Cascading:http://oreil.ly/19Gc9y6
● Cascading website: http://www.cascading.org
● Cascading Lingual website: http://cascading.org/lingual
● Cascading Pattern website: http://www.cascading.org/pattern
● Company: http://www.concurrentinc.com
● Contact Us: http://www.concurrentinc.com/contact
● Follow us on Twitter: http://www.twitter.com/concurrent

About Concurrent, Inc.

Concurrent, Inc.’s vision is to become the #1 software platform choice for Big Data applications. Concurrent builds application infrastructure products that are designed to help enterprises create, deploy, run and manage data processing applications at scale on Apache Hadoop. Concurrent is the mind behind Cascading™, the most widely used and deployed technology for Big Data applications with more than 75,000+ user downloads a month. Used by thousands of data driven businesses including Twitter, eBay, The Climate Corp, and Etsy, Cascading is the de-facto standard in open source application infrastructure technology. Concurrent is headquartered in San Francisco. Visit Concurrent online at http://www.concurrentinc.com.

Media Contact
Danielle Salvato-Earl
Kulesa Faul for Concurrent, Inc.
(650) 340 1982
concurrent@kulesafaul.com

MeetUp | Cascading meetup in Portland – Jul 25, 2013

Event Info – http://www.pdx-hadoop.eventbrite.com

When:
Thu, Jul 25 6:30-9:30 PM (PDT)

Where
Widmer Brothers Brewery – GreatRoom (Gasthaus)
955 North Russell Street
Portland, OR 97227

Organized/Sponsored By: Aaron Betik, NIKE, Inc
Global Technology Director, Consumer and Digital Analytics & BI

Doors are open at 6:15, Talk starts at 7pm. Light snacks, beverages and brews will be available and will have a social hour following the talk by Paco Nathan.

Cascading is an open source workflow abstraction atop Hadoop and other Big Data frameworks, with a 5+ year history of large-scale Enterprise deployments. For example, half of Twitter’s total compute uses this API, along with other large use cases at eBay, Etsy, Airbnb, LinkedIn, Apple, Climate, Nokia, Factual, Telefonica, etc. Cascading leverages some aspects of functional programming so that developers can create large-scale data pipelines which are robust and easier to operationalize. There are popular DSLs in Scala (Scalding) and Clojure (Cascalog), plus Jython, JRuby, etc. Recent support also implements DSLs for ANSI SQL (Lingual) and PMML (Pattern).

Concurrent, Inc. and O’Reilly Media Unveil ‘Enterprise Data Workflows with Cascading’ Book

  • Learn Everything You Need to Know about the Most Widely Used and Deployed Technology
    for Big Data Applications; Online Access Now Available
  • Author Paco Nathan to Speak at OSCON 2013

SAN FRANCISCO – July 10, 2013 – Concurrent, Inc., the enterprise Big Data application platform company, today announced the book release of “Enterprise Data Workflowswith Cascading” by Paco Nathan, director of data science at Concurrent. Published by O’Reilly Media, the hands-on book introduces readers to Cascading™, the most widely used and deployed technology for Big Data applications with more than 75,000+ user downloads a month. The Cascading framework enables users to quickly and easily build powerful data processing applications on Apache Hadoop that span ETL, data preparation and analytics with one unified development framework. The book offers developers, data scientists and system/IT administrators a quick overview on Cascading’s streamlined approach to data processing, data filtering and workflow optimization using sample applications based on Java, ANSI SQL, PMML, Scala and Clojure. Thousands of companies such as Etsy, Razorfish, TeleNav and Twitter already use Cascading for business-critical applications.

With “Enterprise Data Workflows with Cascading,” readers will learn how to:

• Examine best practices for using data science in enterprise-scale applications
• Use workflows beyond MapReduce to integrate to other frameworks and existing IT systems/tools
• Quickly build and test applications, and instantly deploy them onto Apache Hadoop
• Easily discover, model and analyze both unstructured and semi-structured data
• Seamlessly move and scale application deployments from development to production, regardless of cluster location or data size

To purchase a copy of “Enterprise Data Workflows with Cascading” now, visit: http://oreil.ly/19Gc9y6

About the Author

Paco Nathan is director of data science at Concurrent, Inc., where he leads the company’s developer outreach program. He has a dual background from Stanford in mathematics and statistics, and distributed computing. With more than 25 years experience in the technology industry, Nathan is an expert in Hadoop, R, predictive analytics, machine learning and natural language processing. This book release comes on the heels of Concurrent’s recent announcement of Pattern, a free, open source, standard-based scoring engine, built on Cascading, that enables analysts and data scientists to quickly deploy machine-learning applications on Apache Hadoop. Cascading provides the most comprehensive application framework for Hadoop. With the addition of Lingual (ANSI SQL) and Pattern (PMML), Cascading bridges the gap and allows enterprises to use existing skills and systems to easily develop and deploy robust applications on Hadoop. The combination of the three (Java, SQL, PMML) completes the application ensemble.

Learn More About Cascading at OSCON

Nathan will speak on Cascading at O’Reilly OSCON in Portland on July 25, 2013. For more information on his session, “Using Cascalog to Build an App with City of Palo Alto Open Data,” and to register, please visit:http://bit.ly/1azv6oS.

Supporting Resources

● Enterprise Data Workflows with Cascading:http://oreil.ly/19Gc9y6
● Cascading website: http://www.cascading.org
● Company: http://www.concurrentinc.com
● Contact Us: http://www.concurrentinc.com/contact
● Follow us on Twitter: http://www.twitter.com/concurrent

About Concurrent, Inc.

Concurrent, Inc.’s vision is to become the #1 software platform choice for Big Data applications. Concurrent builds application infrastructure products that are designed to help enterprises create, deploy, run and manage data processing applications at scale on Apache Hadoop. Concurrent is the mind behind Cascading™, the most widely used and deployed technology for Big Data applications with more than 75,000+ user downloads a month. Used by thousands of data driven businesses including Twitter, eBay, The Climate Corp, and Etsy, Cascading is the de-facto standard in open source application infrastructure technology. Concurrent is headquartered in San Francisco. Visit Concurrent online at http://www.concurrentinc.com.

Media Contact
Danielle Salvato-Earl
Kulesa Faul for Concurrent, Inc.
(650) 340 1982
concurrent@kulesafaul.com

Concurrent’s Chris Wensel: The Open Source Path Is a Rocky Road

Concurrent’s Chris Wensel: The Open Source Path Is a Rocky Road
By Jack M. Germain LinuxInsider
July 2nd 2013
http://www.linuxinsider.com/story/Concurrents-Chris-Wensel-The-Open-Source-Path-Is-a-Rocky-Road-78398.html#sthash.JnrHmu3b.dpuf

“I want the same clout as Oracle. I just don’t want that same infrastructure as Oracle. Open sourcing is a great way to teach people how to write code, see how things work, and get contributions and get people to trust it. It is marketing as well, however. It is a lot of things. Just one thing is missing: It is not a very clean way to make money.”

Big Data and open source software may be the next great unholy alliance in computing’s current promised land, but open source is a broken business model that needs a better vehicle for supporting projects such as programming suites that build database applications.

So argues Chris Wensel, founder and CTO of Concurrent. Wensel started the company in 2008 to focus on the open-source Cascading Project, a framework that he created to better work with Big Data applications.

Wensel quickly discovered that he was part of a bad news / good news scenario. The bad news was that he began his company too early. The good news is the Big Data market is now growing.

Big Data is the process of collecting and analyzing huge volumes of structured and unstructured data. The data sets are so large that they defy handling with traditional database and software techniques.

Cascading is an open-source framework — Java library and runtime — that enables developers to simply develop rich enterprise-grade data analytics and data management applications. The application can be deployed and managed across a variety of Apache Hadoop computing environments on which database apps typically run.

Cascading was named by InfoWorld as the 2012 Best Open Source Software Winner. Wensel, meanwhile, is also the author of Scale Unlimited Apache Hadoop BootCamp, the first commercial Apache Hadoop training course.

In this interview, LinuxInsider talks to Wensel about the challenges of working with Big Data and the hurdles posed by open source software.

Read the entire interview here – http://www.linuxinsider.com/story/Concurrents-Chris-Wensel-The-Open-Source-Path-Is-a-Rocky-Road-78398.html#sthash.JnrHmu3b.dpuf

Concurrent, Inc. Appoints New VP of Engineering to Lead Product Development

Company Behind Cascading Strengthens Executive Team with Addition of Big Data Veteran

SAN FRANCISCO – July 2, 2013 – Concurrent, Inc., the enterprise Big Data application platform company, today announced that it has named Paul O’Leary as its vice president of engineering. O’Leary, a seasoned veteran, brings more than 20 years of engineering and senior-level experience to the company behind Cascading™, the most widely used and deployed technology for Big Data applications with more than 75,000+ user downloads a month. His appointment strengthens the Concurrent executive leadership as the company scales to support new growth.

As vice president of engineering, O’Leary is responsible for managing the company’s product development. O’Leary most recently served as founder and CTO of Quantivo Corporation, a cloud-based analytics and Big Data management company, which was recently acquired by Aggregate Knowledge. Prior to Quantivo, O’Leary also held various vice president of engineering and senior-level roles at Privia, Inc., Proveer, Kontiki, Broadbase and IBM.

O’Leary’s appointment follows Concurrent’s recent announcement of Pattern, a free, open source, standard-based scoring engine, built on Cascading, that enables analysts and data scientists to quickly deploy machine-learning applications on Apache Hadoop™. Cascading provides the most comprehensive application framework for Hadoop. With the addition of Lingual (ANSI SQL) and Pattern (PMML), Cascading bridges the gap and allows enterprises to use existing skills and systems to easily develop and deploy robust applications on Hadoop. The combination of the three (Java, SQL, PMML) completes the application ensemble.

Supporting Quotes

“Paul’s deep experience in the development and implementation of cutting-edge technology and products is an invaluable asset to Concurrent as we execute and deliver on our core product strategy. Applications are front and center in the Big Data world, and we’re happy to have Paul leading and driving the product development effort and welcome him to the team.” -Gary Nakamura, CEO of Concurrent

“In the last few years, Big Data has exploded across all industries and now, more than ever, enterprises need a reliable and repeatable way to build Big Data applications, quickly and easily. As the brains behind Cascading, Concurrent has already tackled and solved that problem, and is
continuing to drive innovation. I’m excited to join the Concurrent leadership team and look forward to being a part the company’s continued success.” -Paul O’Leary, Vice President of Engineering of Concurrent

About Concurrent, Inc.

Concurrent, Inc.’s vision is to become the #1 software platform choice for Big Data applications. Concurrent builds application infrastructure products that are designed to help enterprises create, deploy, run and manage data processing applications at scale on Apache Hadoop.

Concurrent is the mind behind Cascading™, the most widely used and deployed technology for Big Data applications with more than 75,000+ user downloads a month. Used by thousands of data driven businesses including Twitter, eBay, The Climate Corp and Etsy, Cascading is the defacto standard in open source application infrastructure technology. Concurrent is headquartered in San Francisco. Visit Concurrent online athttp://www.concurrentinc.com.

Media Contact

Danielle Salvato-Earl
Kulesa Faul for Concurrent, Inc.
(650) 340 1982
concurrent@kulesafaul.com