A full house, lots of funny names and what does it all mean?
Cloudera, Appfluent and Informatica partnered today at Informatica World in Las Vegas to deliver together a one day training session on Introduction to Hadoop and Big Data. Technologies overview, best practices, and how to get started were on the agenda. Of course, we needed to start off with a little history. Processing and computing was important in the old days. And, even in the old days it was hard to do and very expensive.
Today it’s all about scalability. What Cloudera does is “Spread the Data and Spread the Processing” with Hadoop optimized for scanning lots of data. It’s the Hadoop File System (HDFS) that slices up the data. It takes a slice of data and then takes another slice. Map Reduce is then used to spread the processing. How does spreading the data and the processing help us with scalability?
When we spread the data and processing we need to index the data. How do we do this? We add the Get Puts. That’s Get a Row, Put a Row. Basically this is what helps us find a row of data easily. The potential for processing millions of rows of data today is more and more a reality for many businesses. Once we can find and process a row of data easily we can focus on our data analysis.
Data Analysis, what’s important to you and your business? Appfluent gives us the map to identify data and workloads to offload and archive to Hadoop. It helps us assess what is not necessary to load into the Data Warehouse. The Data Warehouse today with the exponential growth in volume and types of data will soon cost too much unless we identify what to load and offload.
Informatica has the tools to help you with processing your data. Tools that understand Hadoop and that you already use today. This helps you with a managing these volumes of data in a cost effective way. Add to that the ability to reuse what you have already developed. Now that makes these new tools and technologies exciting.
In this Big Data and Hadoop session, #INFA14, you will learn:
- Common terminologies used in Big Data
- Technologies, tools, and use cases associated with Hadoop
- How to identify and qualify the most appropriate jobs for Hadoop
- Options and best practices for using Hadoop to improve processes and increase efficiency
Live action at Informatica World 2014, May 12 9:00 am – 5:00 pm and updates at:
“Only those who constantly retool themselves stand a chance of staying employed in the years ahead.” – Thomas J. “Tom” Peters
This is a pretty powerful statement. Powerful because it incites an emotional reaction while at the same time forcing a person to step outside of their comfort zone. The emotional reaction comes from hearing the words ‘staying employed’ – as this is something that is important to most of us. Taking us outside of our comfort zone is the self-scrutiny that this quote naturally generates. What are we doing to advance our skills, tools and knowledge to keep pace with what is going on all around us? (more…)
Why performance based testing? Performance based testing measures your ability to apply your knowledge and achieve a specific outcome. Customer success demands that a practitioner’s ability to implement, configure and use our products is validated. (more…)
Informatica customers are faced with many options when selecting vendors and services, and it isn’t simply pricing that drives the decision making process. To be noticed today, data integration professionals need the credentials and reputation that official certification can provide.
Now, more than ever, it comes down to proven qualifications when making a vendor or staffing choice. Certification provides the assurance that the required skills have been demonstrated and validated for project success. A recent study by Constellation Research confirms that “peace of mind” deployments are the number one thing that companies expect from solution providers. Certification brings peace of mind that the project team has the right skills needed to do the job. Enabling our partners and customers to demonstrate their skills objectively is the driving force behind the transformation of Informatica’s new certification program. (more…)
Regardless of your specific industry focus, there is no denying that the tsunami of big data is changing the way we do business. Our customers have admitted that along with the anticipation and excitement of being in the midst of this big technological wave, there is also some anxiety about how to manage all this data. We can practically feel the disciplines of data integration, master data management (MDM), data governance and data quality rising up in importance.
But ask yourself this: Do you have a strategy in place today to ensure that the business has confidence in your data? You may already know that Informatica’s data quality solution helps increase revenue, reduce cost and manage risk. But, do you have the knowledge and skills needed to profile, standardize, match and consolidate data as effectively and efficiently as you can? (more…)
Are you accelerating your company’s project delivery needs?
Many Fortune 500 and Global 500 companies understand that their mission-critical data must be managed, so if you are a developer with the desire to work faster and better you should think about Informatica University Training and Certification.
Why? Because you can establish yourself as a key resource with the skills and experience to 1) ensure a trusted view of your organization’s business data, 2) to accelerate your company’s project delivery needs, or just 3) to demonstrate that you have the needed skills to provide your business the right information at the right time. (more…)