Monday, June 17, 2013

Why You Need To Start Focusing On Social Intelligence

AppId is over the quota

How's your company's social intelligence?

Ask a roomful of business leaders that question, and you're likely to get a variety of different answers, ranging from "Great!" and "Getting better every day!" all the way to "Remind me". What's social intelligence? "

If you're on the latter end of that spectrum and still feel uncertain about how social networks like Twitter, Facebook, YouTube, SlideShare and/or LinkedIn can affect your business, I have to be honest: It's time to up your game. The truth is, social networks like these and many others are affecting your business – and if you're not paying attention, you're not only losing valuable ground to your competitors, your losing invaluable relationships with your customers.

Don't get me wrong. I'm not here to "hype" engagement across social media platforms. And the last thing I want to do is add to the list of problems keeping you up at night.

But, if you're a business leader interested in revenue growth – and who isn't? – you need to start focusing on social intelligence. Every day now, billions of messages are shared across social platforms. How many of them mention your company, product, service, industry, your employees ... or even you? Are those comments glowing recommendations, nasty complaints or something in-between? How many are threats? How many are opportunities? Don't you want to know?  How well are you "listening" to your marketplace?

Companies need to become more attentive and improve their understanding of how social media impacts revenue – both positively and negatively. After all, social media conversations are now shaping the marketplace more – and increasingly faster – than most companies can keep up with.

For example, new social interaction metrics are changing the way value is assigned to television audiences. In other words, the networks aren't necessarily gauging a program's success solely on the basis of audience size. They're also considering the social media activity and habits of certain audiences. As Kevin Glacken explained last week at Smart Data Collective:

"It's understandable that the traditional audience size approach has survived for the past 50 years or so given the one-way channel television has been over that period. However, today with the countless ways for viewers to immediately react to and interact with their television programs, quality of viewers in terms of interaction and insight is matching sheer audience volume in many cases. "

"Programs that have high IVRs (Interactive Viewer Ratings, a metric developed by ListenLogic) can gather to deep, multidimensional understanding of their audience members via advanced social intelligence in terms of interests, activities, likes, dislikes, attitudes and behaviors. This ultimately can deliver better targeted and effective messaging to advertisers. This increases the efficacy, and thus the value of the program. "

At Teradata, we know it's increasingly essential for all companies – whether they're B2C or B2B – to tap into the insights available from social media big data streams. Even better, we know that true business value emerges when social intelligence is integrated with other internal practices. Put another way:

Big date scope + freshest delivery date + behavioral and contextual data + comprehensive predictive analytics + real-time messaging = Maximized business results.

With its thorough and detailed data-driven understanding of customer behaviors, Teradata's Interactive Customer Engagement now empowers marketers to:

Combine online and offline data to reveal new insights, Deliver optimized, personalized real-time offers, based on historical and contextual in-session data and search results;Offers coordinates across online and offline channels in real-time;Leverage industry-leading campaign management and email delivery functionality to identify and communicate more relevant and personalized offers to the customer;Gain a clearer view of both the customer's path to purchase as well as his or her journey through various channels;Have a better understanding of paid channels ' roles and effectiveness in converting browsers into buyers; Deliver relevant, meaningful, real-time offers across multiple channels by learning from previous interactions and self-aligning with company goals.

The big data generated by streaming social media offers a wealth of information about your brand, your industry, your competition and most important of all, your customers. It's time to start listening. It's time to start understanding. And it's time to start putting all this information to work for you.

-Darryl


Defining the Appliance Marketplace

AppId is over the quota

Abbiamo avuto recentemente l'opportunità di parlare con Roxanne Hendricks, Product Manager di Teradata elettrodomestici nuovi Teradata ® Data Warehouse Appliance 2700. Questa premier piattaforma offre prestazioni migliorate e migliorata gestione dei workload.

Hendricks sottolinea, nel suo Q&A, che il mercato è cambiato significativamente poiché Teradata ha introdotto il suo primo apparecchio nel 2008. Negli ultimi cinque anni, le imprese sono venuti a fare affidamento su apparecchi di Teradata per disaster recovery, piattaforme ibrido, sabbiere di sviluppo per Google analytics e anche per integrare dati integrata magazzini. Non importa l'impiego, organizzazioni non solo guadagno rapido accesso alle informazioni, ma anche il tempo più veloce di valore.

Brett Martin
Senior Editor
Teradata Magazine


EPIC Awards 2013 Teradata – nominate your company

Okay, this probably isn't new to you: Teradata-we love to data and all the amazing things you can do with it. Analysis and data storage is a big thing for almost all companies worldwide. And while many are already doing quite well, some do it even better. A crisp example is McCain Foods Limited. The Canadian producer of oven-ready frozen products with numerous facilities and offices around the world is able to provide all its employees not only with data but with intelligent and actionable information acquired by advanced data analysis – local as well as on corporate level. In this way, the data company departments and employees in each site no longer need to gather and extract data to provide to their clients or colleagues, because they have direct access to the information they need through reports and dashboards.

Since Teradata is so excited about data analysis, we are always happy to hear about these success stories from our customers and, of course, also from our partners who help make this happen. Therefore we honored McCain and 14 other companies with the epic Awards 2012 Teradata. And since we have not changed our opinion last year, now we're ringing the Bell for the 2013 Edition. The winners can be significantly enhanced that Teradata customers their business objectives by implementing analytical solutions best-in-class, or Teradata partners companies such as independent Software vendors (ISVs) and system integrators (SIs) that outstanding contributions to Teradata and its customers, and who have demonstrated their commitment to the success of business through the Teradata platform.

So if your team, too, has created a Teradata-based solution that deserves international recognition, appoint the company until July 19, 2013. Winners will be announced at the awards ceremony of Teradata EPIC held during Teradata Partners Conference, User Group Expo & this October in Dallas. Look forward to many great and inspiring entries!


5 steps to making BI more intelligent in big data Analytics

Translate Request has too much data
Parameter name: request
AppId is over the quota

In recent months, I met with the Business Intelligence (BI) teams in different countries to discuss Big Data Analytics. What transpired from the meetings is clear lack of awareness of what Big Data Analytics can do for the BI team and how Big Data Analytics fit within the enterprise data warehousing (EDW). As ambassadors to their business community, BI teams have the opportunity to be at the forefront of new technology trends and be able to articulate the value of Big Data Analytics to business stakeholders.

The Big Data trend has been here for a while and there is no shortage of publically available resources on the subject. However, many of these sources do not seem to allow the audience “to see the wood for the trees”! Also, storage vendors such as Dell and EMC are not helping the situation either by confusing the BI teams with low cost storage aspects in preference over business value of Big Data Analytics. I believe that paying attention to business value of Big Data Analytics will make the BI team not only look smarter in front of the business stakeholders but also make it easier to get funding for Big Data Analytics projects which many of the BI teams are considering as an opportunity to advance their career ambition.  

In the next few paragraphs below I have described in a few steps some essentials of Big Data Analytics in technical terms and how they fit into the enterprise data warehousing ecosystem as unified data architecture (UDA) that supports the next era of analytics and business insights. Many of these examples are related to the airline industry but the principles equally apply  to any industry.   

Step 1: Getting to know the essentials of Big Data

First step to Big Data Analytics is to understand the new technology capabilities such as Map Reduce, Hadoop, SQL-Map Reduce (SQL-MR) and how they fit within the enterprise ecosystem. It is also important to understand the differences in approaches between traditional EDW and Big Data Analytics design, development and implementation processes.

For instance, if you are in the airline industry, you would have designed the enterprise data warehouse for transactional reporting and analysis with structured stable schema and normalised data model.

You probably stored unstructured data such as ticket image, recorded audio conversations with customer service agent and ticketing / fare rules in the database as BLOB (Binary Long Object). Furthermore, you may have found it difficult to write in declarative SQL language the complex business rules such as financial settlements of inter-line agreement from code share arrangements, open jaw fare rules, say between Zone 1 and Zone 3, and business rules for fuel optimisation; so, you may have resorted to procedural languages such as user defined functions (UDF).

But UDFs have numerous limitations that MapReduce, more specifically, SQL-MapReduce (SQL-MR) makes it easy to solve while allowing for high performance parallel processing.

- What if you are able to use MapReduce API (Application Programming Interface) through which you can implement a UDF in the language of your choice?
- What if this approach allows maximum flexibility through polymorphism by dynamically allowing determination of input and output schema at query plan-time based on available information?  
- What if it increases reusability by enabling inputs with many different schemas or with different user-specified parameters?
- Further, what if, SQL-MR functions can be leveraged by any BI tools that you are familiar with?

As you can guess, SQL-MapReduce (SQL-MR) overcomes the limitations of UDF by leveraging the power of SQL to enable Big Data Analytics by performing relational operations efficiently while leaving non-relational tasks to procedural MapReduce functions.

You will see some examples of this later but, first and foremost, what is MapReduce? MapReduce is a parallel programming framework invented by Google and popularised by Yahoo!.MapReduce enables parallelism for non-relational data. By making parallel programming easier, MapReduce creates a new category of tools that allows BI teams to tackle Big Data problems that were previously challenging to implement. It should be noted that unlike the core competency for parallelism of the Teradata’s relational database technology over the last 30 years, MapReduce is not a database technology. Instead, MapReduce relies on file system called Hadoop Distributed File System (HDFS). Both MapReduce and HDFS are the open source versions of the Big Data technologies.

Step 2: “Hello World” welcomes you to the world of MapReduce with “Word Count”

Let’s take look at how Hadoop MapReduce works! When you wrote your first program you may have tested it to make sure “Hello World” works by printing / displaying the words correctly. With MapReduce, you will most likely to be testing Word Counts in your MapReduce program.

A MapReduce (MR) program essentially performs a group-by-aggregation in parallel over a cluster of machines. A programmer provides a map function that dictates how the grouping is performed, and a reduce function that performs the aggregation.

Let’s say that you want to create a Book Index from Big Data Analytics for Dummies. When writing your MR program, you will provide a map function that dictates how the grouping is performed on paragraphs containing words, and a reduce function that performs the aggregation of the words to produce the book index. The MapReduce framework will assume responsibility to distribute the Map program to the cluster nodes where parts of the book is located, processed, and output to intermediate files.  The output of the map processing phase is a collection of key-value pairs written to intermediate flat files. The output of the reduce phase is a collection of smaller files containing summarized data. The key-value pairs of words above are reduced to aggregates that produce the book index.

Because the MR program runs in parallel you will notice tremendous increase in reading (e.g. grouping of paragraphs from Big Data Analytics for Dummies) and processing speed (e.g. summarising and aggregation of key-value pairs) that would impress even Johnny 5

Creating an index list of words and counts from Big Data Analytics for Dummies may not be terribly interesting or useful for you but, the capability of such key-value pair generation from any multi-structured data sources can be put to analytical use by creating a set of useful dimensions and measures that the BI teams are familiar with that can be integrated with data in the EDW. Perhaps, instead of creating the Book Index, you may choose to create an index of all flight numbers, origins and destinations from the booklet of an airline time table which you may find more useful in the airline business.

Step 3: Putting MapReduce to solve business problems

Long gone are the days of GSA’s (General Sales Agents) enjoying hefty sales commissions from the airlines! The market is highly competitive and organisations are looking for best decision possible from analytics. With ubiquitous availability and convenience offered by broadband connections, customers’ attitudes and behaviours are rapidly changing. Now customers are looking for best travel and holiday packages online. They are also listening to the opinions of their friends and public remarks on social network forums. Interestingly, this is also instrumental in rapid rate at which huge volumes of data is generated, opening up the need for Big Data technologies.

What if we could utilise the multi-structured data from click streams, Facebook, Twitter data for improving business performance? What if we are able to extract the IP Address from the click stream data and correlate with the profile of the customer from EDW along with best fare for the Round The World Travel deal that the customer is looking for? What if we are able to extract the sentiment of the customer’s travel experience from Twitter and Facebook data and use the positive / negative experience to provide the Next Best Offer during the customer’s next inbound call to the agent or online visit?

Step 4: Integrating unstructured and structured data for Big Data Analytics

Here we consider how the integration of multi-structured data in MapReduce and structured data in EDW can be used for improving business outcome. You will see that instead of the MapReduce program for Word Count that you wrote previously, you will write a new MapReduce program to extract the key-value pairs for IP Address, flight deals and any other relevant information from the Apache Weblog files where the customer’s online interaction is recorded. In a later paragraph I will describe how the MapReduce program you wrote is invoked in SQL by means of SQL-MR or better still how you can leverage several pre-built functions (without having to write your own MapReduce program) using SQL-MR. For now, let’s assume the extracted data from MapReduce is created as a table in the EDW. The extracted IP Address can then be joined with Master Reference in the EDW to identify the User ID which is then used to match the frequency of online visits and lifetime value of the customer etc.

Step 5: Flying high with SQL-MR (SQL-MapReduce)!

While MapReduce is good for solving Big Data problems it can cause a number of bottlenecks, including the requirements to write software for answering new business questions. Trying to exploit data from HDFS through Apache Hive is another story; let’s not even go there! SQL-MapReduce (SQL-MR) on the other hand helps to reduce the bottleneck of MapReduce by allowing maximum flexibility through polymorphism (by dynamically allowing determination of input and output schema at query plan-time based on available information). It allows reusability by enabling inputs with many different schemas or with different user-specified parameters. More importantly, you can exploit all types of Big Data using the BI tools that you and your business analysts are familiar with.

Here you will see examples of how you may use the SQL-MR function text_parser (with just a few lines of code) to solve the word count problem / creation of a Book Index for Big Data Analytics for Dummies / extraction of IP Addresses from online clickstream data. You will notice reusability of the SQL-MR function that enables inputs with many different schemas and with different user-specified parameters to create output schema at query time.

You will find that SQL-MapReduce (SQL-MR) provides excellent framework for jump starting Big Data Analytics projects with substantial benefits, viz. 3 times faster in development efficiencies, 5 times faster in discovery and 35 times faster with analytics. My colleague, Ross Farrelly, demonstrates with an example of how to reduce the pain of MapReduce ,which will be of interest to you as well. You can see how SQL-MR provides an excellent framework for customising / developing SQL-MR functions easily with an Integrated Development Environment (IDE).

Exploring and discovering value from Big Data is how you will divide and conquer the volume, velocity, variety and complexity characteristics of Big Data. You will also gain great benefits from seamless integration of the different Big Data technologies as a Unified Data Architecture (UDA) to provide advanced analytics.

Here is another business use case that the SQL-MR functions nPath and GraphGen solve elegantly and efficiently compared to either SQL or MapReduce. Try writing this in SQL or MapReduce and notice the difference! The business problem that we are trying to solve is related to identifying the more frequent customer activities or sequence of events that lead to disloyalty.

You can see from the chart below that of all the different channels that customers use to buy airline tickets, the online channel leads to unsuccessful ticket sale. By visualising the sequence of all customer events you will notice that the Online Payment page is where abandonment occurs (i.e. noticeable from the thick purple curved line that indicates the strength of the path segment) which provides insights about the issues with the online channel. By taking corrective actions ahead of the online payment event step you will create customer loyalty and growth in sales.      

Here is the SQL-MR code for the above visualisation of ticket purchase path analysis:

If you are all set and ready to go on your first class journey with Big Data Analytics then, check-in here .While ‘inflight’, treat yourself with ‘cocktail’ of analytical functions from a wide ranging selection of 70+ pre-built SQL-MR functions .

Travel smart, impress your accompanying business stakeholder, double your rewards from analytical outcomes and enjoy your journey with Big Data Analytics! By the way, don’t forget to drop me a note, if you found this useful! Bon voyage!

Sundara Raman is a Senior Communications Industry Consultant at Teradata ANZ. He has 30 years of experience in the telecommunications industry that spans fixed line, mobile, broadband and Pay TV sectors. At Teradata, Sundara specialises in Business Value Consulting and business intelligence solutions for communication service providers.


Five Ways Big Summer Date Will Impact Travel

AppId is over the quota

Memorial Day è proprio dietro l'angolo e con esso arriva l'inizio ufficiale dell'estate. Per la maggior parte di noi – se sei sette o settanta-anni – la parola "estate" è sinonima di "vacanza", e le vacanze, naturalmente, coinvolgono viaggi.

Infatti, secondo recenti risultati dell'indagine da TripAdvisor, circa un terzo (30 per cento) degli Stati Uniti intervistati intervistati stanno progettando di viaggio questo Memorial Day weekend. Quasi nove su dieci (86 per cento) stanno progettando un viaggio di piacere quest'estate. Che è un aumento del 6% e il 7 per cento, rispettivamente, rispetto ai risultati dell'indagine di TripAdvisor dall'anno scorso.

Che cosa viene in mente quando si pensa a viaggi estate? Si tratta di spiagge? Sight-seeing? Parchi nazionali, forse?

Non io. Forse non venire come una sorpresa per chi mi conosce bene, ma quando penso a un viaggio in questi giorni, in genere finisco per ruminanti su una cosa: grandi quantità di dati... o, più specificamente, grandi quantità di dati analytics e come piattaforme e applicazioni di oggi stanno facendo viaggiare più facilmente in tutto il mondo.

Proprio come ogni altro settore, i viaggi e turismo industria ora affronta la sfida di acquisizione, memorizzazione e trovare valore in enormi quantità di dati. Comprensioni di affari – quando sono velocemente accessibili e costruito sull'infrastruttura dati destra – pu? trasformare gut-sentire le decisioni in quelli intelligenti, basate sui fatti.

Presso Teradata, sappiamo che quando le imprese di viaggi e turismo sceglie di integrare i processi operativi e strategici di business e intelligenza sociale, iniziano a guida soddisfazione del consumatore e la redditività. Perché? Perché le intuizioni di grandi quantità di dati direttamente l'impatto:

i costi di viaggio. Compagnie aeree, agenzie di viaggi, Alberghi e relativi servizi operano in un contesto iper-competitivo dove rapide risposte alla domanda dei consumatori, concorrenza e condizioni ambientali possono fare la differenza tra i profitti e le perdite nelle operazioni.

Disponibilità di biglietti.Come con i prezzi dei biglietti, attento monitoraggio della domanda dei consumatori, concorrenza e le condizioni ambientali consentono viaggi e turismo le imprese ad adattarsi a un mercato liquido.

L' esperienza del cliente. Intuizioni di grandi quantità di dati guida viaggi e turismo le imprese migliorano l'esperienza del cliente offrendo spunti critici per le linee del fronte solo quando necessario – anche in tempo reale.

Lealtà. Viaggi e turismo le imprese che si avvalgono di intuizioni di grandi quantità di dati costruire fedeltà con offerte progettate da particolari "nascosti" nei dati del cliente.

Un vantaggio competitivo.Imprese viaggi e turismo utilizzano grandi quantità di dati analytics per trovare i driver di riduzione di costo in operazioni irregolari, l'utilizzo di produttività, carburante e materiali di lavoro, ecc.

Vuole un esempio? Basta guardare a come colleghi di eCircle Teradata utilizzato dati e applicazioni per mantenere Londra spostando l'estate scorsa durante il 2012 Olimpiadi e Paraolimpiadi. A nome del suo cliente Transport for London (TfL), Teradata eCircle implementato selezione dati sofisticati per creare email mirate che potrebbero essere utilizzate per comunicare direttamente con i pendolari quotidiani, inviando loro un sito speciale, dove essi potevano piano loro viaggi per evitare Viaggi hotspot. A seguito di questi messaggi di posta elettronica, 35 per cento dei londinesi ha apportato modifiche alle loro rotte, riduzione del traffico nel centro di Londra da circa il 15% durante i giochi.

Mi aspetto analisi di grandi quantità di dati per avere un impatto sempre più significativo sul viaggio nel prossimo futuro. Secondo il World Travel & turismo Consiglio (WTTC), l'industria dei viaggi e del turismo dovrebbe crescere in media del 4 per cento all'anno nel prossimo decennio. Notevolmente, che sarà pari al 10 per cento del PIL mondiale, o circa US$ 10 trilioni. Il WTTC anticipa anche che da 2022, viaggi e turismo rappresenterà per 328 milioni posti di lavoro, o in ogni 10 posti sul pianeta. Con tutte queste persone "on the go" – e tutte le informazioni digitali che siete tenuti a generare – rende perfetto senso per le imprese di viaggi e turismo iniziare sfruttando Google analytics di grandi quantità di dati per migliorare l'esperienza del cliente... e non solo per il Memorial Day o le vacanze estive, ma per ogni giorno dell'anno.

-Darryl


See the big picture

Apache Hadoop ™ ® has gained popularity for its ability to process large data sets using the simple programming. As part of Teradata ® Unified Data Architecture ™, Hadoop lays the foundations of analytical data to quickly and easily store and access large volumes of data.

In a nutshell, Arlene Zaima, program manager for Teradata integrated Analytics solutions, says that Hadoop data contain "hidden jewels" that can be integrated with the information in the data warehouse to learn more. Teradata did find those gems easier in two ways. Smart charger allows bi-directional data transfer between Hadoop and Teradata and a new interface named Teradata SQL query-H ™ helps analysts use standard interfaces to access data in Hadoop.

Enabled by this self-service environment, analysts have the luxury to enrich information from an integrated data warehouse with highly valuable data from Hadoop to find new business value, from large data analysis.

Brett Martin
Senior Editor
Teradata Magazine


Volare Airlines superiore con Google Analytics

AppId is over the quota

Airlines, plagued with sluggish profits and frustrated customers, realize the growing importance of data analytics as customers vent their frustration across the social media universe. Dr. Nawal Taneja, an aviation industry expert and author, says carriers are facing both a challenge and an opportunity to create new customer experiences and increase profitability. In response, airlines are working ever harder to not only to better understand what makes happy and loyal passengers but also which ones are the most profitable.

The solution is better use of data and technology, such as data warehousing and analytics, that allows companies to transition from being service providers to providing solutions; to shift from a fee-based to a value-based business model. Now when they look toward the end of the runway, industry leaders are getting a glimpse of a future in which they'll be able to choose passengers, rather than the other way around.

Brett Martin
Senior Editor
Teradata Magazine


From the countryside of commitment: Teradata helps CPG marketing as connect with today's consumer

Consumer Centricity is the new "black" between a retailer of consumer goods.  Knowing who is buying your products when and where – the purchase path – is fast becoming table stakes for marketing competing for the attention of walletshare and poor consumers.  It is an easy task, considering the diversity of points of contact and data-dependencies illustrated by Don Scheibenreif Gartner Customer 360 Summit recently:

 

 Consumer Insight

Sometimes we start from the ruler the silo and often outside data consumer generated by numerous agencies employed by digital, brand and shopper marketing.  This data – and analysis of data derived from it – can inform intra-and cross-brand marketing aligned to key decision points along the path to purchase, through channels such as the web, email, social media and mobile.  Is a powerful concept that we talked recently to the Teradata Conference APEX, and relies on technology management leadership and commitment.

This requires not only a change in the mentality around the consumer data (which is good), it takes a commitment to ongoing commitment of consumers that differs from seasonal and promotional marketing-driven.  This type of dialog with consumers launches the switch on the product-driven messages to those that instead of serving the consumer so friendly and considerate.  Doing this well depends on a living record of every consumer and shopper who does business with your brand.

Before embracing this approach, some struggle with a perceived disconnect between marketing budgets shrinking or static and the need to increase the level of interaction with consumers and buyers.  "Out of sight out of mind" never applied more than brands that today given the saturation of messages aimed at consumers online or in-store.  Thus it requires a real commitment (Executive) for the commitment (albeit respectful permissions preferences and communications).

Marketing Resource Management

For this reason, other consumer goods organizations approach this opportunity operationally focused on the efficiency of the marketing organization, people and processes.  Marketing Resource Management or MRM, is more easily explainable as an ERP system for marketing ".  Garter defines as:

"... .a set of processes and features designed to improve a company's ability to orchestrate and optimize resources, internal and external marketing. "

As the functionality for data management and engagement, Teradata solution for MRM is also ruling class (the MRM Magic Quadrant shown here).

The three key functional areas of MRM – spend management, workflow and asset management-bring automation, supervision and simplification for burdensome manual processes involved in working with and paying agencies, develop and launch campaigns and leveraging the content and activities developed over time by the entire marketing organization.  The savings can be significant – as in tens of millions.

Increased efficiency leads to newfound agility.  Understand what works and what doesn't-faster than before-then deploy more resources for winning marketing programs help to adapt to the digital world real time consumers operate today.

More efficient resource allocation supports investment in a continuous engagement strategy aimed at buying test, repeat and referral marketing higher performance.  If these efforts can still be mapped to consumer and unified profile data from shopper, minus the MRM facilitates improved performance within a brand or a marketing channel as an integrated strategy data comes online.

Ultimately, the campaigns are no longer defined or constrained by long internal processes.  Instead they become engagement engines create permanent consumer connections that inform product development, marketing, sales and distribution and supply.  That is the data driven marketing future that all consumer products companies should aspire to.

Bassett GiB


Perché le organizzazioni dovrebbero avere una capacità di individuazione dei dati

AppId is over the quota
AppId is over the quota

I have been struggling to reconcile two different thoughts over the last few months and watching a video recently forced me to think about this again. There seems to be a catch 22 between finding value in new data and having the tools and mechanisms justified by the value to find the value. Firstly I see a lot of organisations struggling to get their analytics initiatives underway and sustainable, there are many articles on the web about this. The second is why should organisations have a data discovery capability, is this a marketing term or is there real value in it?

Something occurred to me recently. We were discussing the value chain of analytics, or something analogous to a value chain for analytics in large organisations and exploring how different pieces of data have different value and how this could be used in a BI Centre of Excellence to engage with users. The question was how do you decide when to put new data into the warehouse and what data remains outside the warehouse?

What occurred to me was that a lot of the value of much of the data we were considering had already been established or simply assumed (someone asked “loudly” enough for it and they got it). It was being stored and managed in a data warehouse, it was being accessed by users using various toolsets and although critical to the management of the business is fundamentally operational in nature. The value of a particular piece of data had been established previously and subsequently significant investment had gone into that piece of data to get it into the warehouse on an ongoing basis. That piece of data was now being used to manage and change the business so it was creating impact.

A short digression - because the value of data is determined by what it can be used to accomplish. Data has no intrinsic value, in fact even insight or actionable insight has no value unless it is put into action or changes something. Unless you change the business, a process or an offering in some way the data is merely interesting not important. Analytics teams are often cut off from the business and the ability to impact the business in a meaningful way.

The question became how do you get new pieces of data added to the operational store, the warehouse, because that is how it gets used and therefore that’s how it becomes valuable when you don’t know it is valuable yet? You have to know it is worth something before you integrate it into the data warehouse because there is a significant investment in integrating a piece of data into the warehouse. You have to know it is worth something before you invest in operationalising it. Seems a little confusing because you have to know it is valuable before you make it available to make it valuable, in which case you don’t actually know it is valuable.

Knowing that a piece of data is worth something is also important in justifying analytics teams and getting analytics initiatives up and going. Until you know something has value for a fact, there is no way to build a business case that funds the building of an analytics team. I have seen a number of organisations try address this problem by employing a small team, sometimes an individual, to be the analytics team. Problem solved as it makes it an opex problem, something the business can fund month to month. But this approach struggles to gain momentum, struggles to justify value and is very difficult to build a long term business case around.

These teams are either accommodated in IT where they can often get the data but not the business question or they are housed in the business and struggle to get the data. In addition to this much of the value in analytics comes from looking across the business. Silos tend to be quite good at optimising their narrow world, the value comes from optimising across silos.

This is when a discovery platform may be valuable. If you can provide easily accessible analytical ‘sandboxes’ that are both easy to use and can access all types of data you change the problem from being one of funding to one of testing the findings. Currently discovering the value in a piece of data is hard. There is no single technology that addresses all requirements requiring users to employ multiple tools. HDFS and Hadoop is attracting a lot of interest but is not the easiest to use, especially for business users. SQL is positioned as more of a business language but does not access all data structures. So what do you do if you want to find valuable data but are skills constrained?

Someone mentioned to me one of the ways this can be done is using an agile analytics methodology or approach. In my experience of agile, admittedly mostly in software development, “agile” has often become an excuse for no documentation, no objective or no accountability so I have been a little sceptical about anything labelled ‘agile’. Admittedly it has come a long way since I first bumped into agile so decided to test my bias and looked through some articles on an internal website about an analytical agility capability. Don’t get me wrong, I buy the drivers for agility – very short term deliverables, direct business involvement, the output is more important than the governance or methodology so I would like it to work.  This was also about analytic agility not so much an agile methodology.

The only way to identify new items of valuable data is to experiment and test. Something I have been hearing is “fail fast” which sounds bad but really means test lots of things, do it properly and determine which ones are not going to work fast. Take successful experiments and operationalise them fast. This is what a discovery platform can enable. I would like to get other peoples view but this to me is a way to rapidly test things and determine which data long term should be integrated into the data warehouse.

There is the emergence of the Discovery platform, a set of technologies that makes it easier to integrate multiple sources and types of data while providing a uniform mechanism to access them. Namely SQL. They also provide a means to test the insight in a rapid way and thereby prove value before investing in operationalising an insight. You get to test the value in a meaningful way and evaluate the value before having to invest in making it available.

If anyone has a view on either an agile approach to identifying valuable data or how a Discovery platform can help in opreationalising analytics it would be great to hear some views.

Craig Rodger is a senior Pre-sales Consultant with Teradata ANZ focusing on advanced analytics. He has spent 20 years in the IT industry working on how to get value out of systems rather than getting things into them. Having been a member of a number of executive management teams in software, technology and consulting companies and helping build a number of technology business ventures he joined an advanced analytics vendor.


La realtà di grandi quantità di dati Analytics

AppId is over the quota

"We tend to overestimate the effect of a technology in the short run and underestimate the effect in the long run," said Roy Amara, former president of the California-based Institute for the Future. How right he was. Think about the old days of the World Wide Web. During the 1990s it was widely believed that the Internet would revolutionize our economy suddenly. And related technological advance was expected to boost profits in the near future. This "irrational exuberance" (Alan Greenspan) then led to the dot-com bubble at the turn of the century. Meanwhile, the Web has changed business dramatically and those Internet-based corporations that managed to survive the turmoil began to see profits or even rose to become industry-dominating enterprises.

The entire story is what Gartner calls a hype cycle. The consulting company has been characterizing what typically happens to new technologies since 1995. And as I recently mentioned, the phenomenon of big data – according to Gartner's Report for 2012 – is about to reach its top of the hype-graph. If that turned out to be true, big date would soon slip towards the trough of disillusionment. Bang! You wouldn't hear much about big data technologies for a while. They would only slowly recover and would finally level at a normal stage, embedded within many other established technologies.

Newsflash! A recent report published by the Business Application Research Center (BARC) in Würzburg, Germany, shows you big date is already transforming companies. The Big Data survey was conducted in major European software markets and answered by 274 decision makers in IT and other departments. The answers show that – beyond all the talk about it – big date has become a reality within European companies, helping managers to gain deep insights into markets and customer needs.

While 14 percent of the companies have already developed a detailed big data strategy, 75 percent are aware of the new possibilities arising from big data technologies. However, most companies still face serious challenges in monetizing big date. Lack of expertise is one of the main obstacles, but you can clearly see the trend: Big date has long since moved into the slope of enlightenment. We are there to help a broad range of users benefit from big data technologies. And I feel we are getting very close to the final stage.


The next episode: Dr. Dre meets large amounts of data

AppId is over the quota

Combattere!

Come un allume USC, è strano che I cringe quando altri Trojan ben intenzionati strillare il grido di battaglia della scuola.  Ma, la settimana scorsa, invece di sentire la retorica di una marching band nella mia testa al pensiero di ' SC ' combattere la canzone, stavo Feelin ' un po ' più hip-hop. Ho avuto uno spirito affine in Dr. Dre.

Cuz che (come direbbe Dre) famosa star hip-hop Dr. Dre e magnate musica Jimmy Iovine ha annunciato una donazione di $70 milioni per l'Università per creare una nuova Accademia per musica, concentrandosi sull'intersezione tra arte, tecnologia, business e innovazione. Il curriculum include Computer Science. Imprenditorialità. Art. Di marketing. Google Analytics.

Tutti noi non potevamo stare a trovare questi incroci un po' più chiaramente?

Questa è la sfida per l'industria media e dell'intrattenimento di oggi – la necessità di trovare tale intersezione di "gut" e "intuizione".  Sono sicuro che questa è la rovina di qualsiasi industria creativa lunga in data odierna driven clima. Ecco perché è difficile contestare la saggezza collettiva di potenze creative che sono state a loro commerci per decenni. Ma, nessuno discute che ci deve essere un interruttore all'ingrosso. Piuttosto, solo alcuni più apprezzamento per l'intersezione.

In una manciata di anni che ho lavorato in analytics – che sono stati preceduti da molti manciate di anni di lavoro in produzione, post-produzione, e digital media – ho visto un vero ramp-up nel ruolo di analytics presso società di media tradizionali e digitali allo stesso modo.  Ma la verità è che ci sono ancora le fazioni.  Se sto parlando a un creatore di contenuti, distributore, editore o MSO, ci sono spesso nei campi: il campo di Google analytics-sono-sopravvalutato contro l'accampamento di Google analytics sono il nostro futuro.  Quei due campi stanno iniziando a incontrarsi nel mezzo – ed è giunto il momento.

A rischio di sembrare eccessivamente profetico, c'è bellezza nell'intersezione tra arte e scienza.  E, che, credo, è la promessa di Google analytics grandi quantità di dati attraverso la catena del valore contenuto. Quando le aziende creative possono integrare ci? che sanno sul loro pubblico, loro contenuto, i canali e loro commercializzazione, si pu? scatenare il valore di intersezione tra arte e scienza.  Qualsiasi quadro google analytics di successo richiede una comprensione dettagliata dell'arte di entrambi.

Cos?, per tutti voi aspiranti artigiani dati là fuori, prendere il cuore! Dr. Dre ha ottenuto la schiena su questo su questo. Combattere!


Seeking: scientist

AppId is over the quota

Un mio amico ha un figlio adolescente che ama computer, videogiochi ed è un esperto di matematica. Quale percorso di carriera di mio amico raccomandando a lui? Scienza di dati! L'adolescente è già in classi statistiche supplementari per ottenere pronto per il college. Speriamo non ci sono più bambini come lui là fuori, perché stiamo andando ad avere bisogno di loro.

Nell'attuale ambiente trovando un dati scienziato pu? essere una lotta. La carenza è radicata nel fatto che solo un gruppo relativamente piccolo di persone eccellere a entrambi i dati avanzati concetti di Google analytics e business che serve per aiutare le imprese a trasformare enormi volumi di dati in intuizioni impugnabile. Nella nostra infografica "Un lavoro ben fatto," Noi esploriamo quello che dicono le parti interessate di trovare questi esperti critici.

E no, Siamo spiacenti, figlio del mio amico non accetta offerte — ancora.

Jennifer Niemel?
Executive Editor
Teradata Magazine


Tips for creating better communication via email

So you've received an email from your boss with the comment "FYI" at the top of the email trail with a subject even more obscure "catch-up with".  So he dutifully goes through the email trail trying to understand the context and significance of why your boss I wanted to read it.  Go through the attachments and try to extrapolate this is relevant and then try to deduce what is exactly the action required of you.  5-10 minutes later you have decided you have read enough and if it were that important, someone will eventually ask you to do something.

Do you feel uncertain.  So you keep a copy of the email, or tag, or put a "read later" folder just in case.  Now multiply this effort with hundreds of emails that you will receive every day and you can see how fast you eat your day.  What is worse, is a vicious circle-forward email to your colleagues and subordinates with his own formulation "FYI".   This is normal lazy and wasteful communication.

If you have already invested time and effort to analyze the communication you received, it would be more productive for the people from going through the same pain of replacement?  Wouldn't it be good to have a simple test that can be used to verify the effectiveness of your communication?  My answer- S C A Q *.  (* derived from the framework communication http://www.barbaraminto.com/by Barbara Minto.)

The acronym stands for situation, complication, question and answer.  Each piece of communication they prepare, if an email, a presentation or a paper proposal should have this contained in the first two or three paragraphs.

So here's how to apply it.:

Situation: my communication sets out the situation clearly so people understand if you apply them or not. This helps filter the people concerned on the issue, saving time and effort.

(S) XYZ SW version 4.3 has been delayed.

Complication: has my communication transmitted what has changed in the situation that creates a sense of urgency for people to pay more attention and be prepared to take some action tomorrow? (C) members involved in pre-sale moto can offer functionality will not be available.

Question: I have framed my question that there is no ambiguity as to what the player should be looking? This prevents the player from second guessing what should be the questions. It also helps to validate the completeness of response that should provide.   (Q) what are the new features in the new release and what is the revised release date?

Answer: you have already framed the issue then you must provide the complete answer to this question. This can be as long as need be.  (A) time and columns are the key new features of the new version. The release date is now Q4 2014.

The entire email now reads:

4.3 SW version XYZ has been delayed. Employees involved in pre-sale moto may be offering functionality will not be available.

What are the new features in the new version and what is the release date changed? And Time columns are the new key features of the new version. The release date is now Q4 2014.  For more details, please read the attached document.

Compares this short but pointless:

Kind regards. Refer to attached.

Now take a look at an email that they were going to send and make a "SCQA" on it!  Send me your before and after release.  Happy to talk at all.

Renato Manongdo is a consultant of Teradata sector ANZ with extensive experience in customer service in financial services, insurance and healthcare industries.


Open Access to Big Data a Major Driver of Value for eBay

AppId is over the quota

At Teradata's Big Data Analytics Summit, held recently in Sydney, Alex Liang of eBay (Director of the offshore Analytics Platform and Delivery) presented on their big data ecosystem. It should be noted that his description has to be taken in the context of eBay – a company in which their business is their website is a marketplace on which they aim to match customer's desires with seller's products. This takes place on a massive scale with a requirement for 99.9 + percent availability.

Nevertheless, despite this challenging environment, eBay is committed to the democratisation of data – that is, making data available to large number of employees to query, predict and experiment. To this end they have a decentralised data management system which allows employees to create a virtual datamart, ingest data, identify a trend or gain an insight, form an hypothesis, design an experiment to test that hypothesis, implement that experiment on the eBay website (via A/B testing), measure the results and undo the changes if necessary – all with a great deal of autonomy.  However, with freedom comes responsibility, so employees using data in this way are also responsible for the results they generate from their data analyses.

Can this approach be applied to other companies? As a general principle it can – the philosophy of allowed a larger number of users access to the valuable data held by company can, if implemented well, lead to outstanding results. The approach of allowing users to run free on the date (within certain well defined limits of course) only reigning them in when they approach those limits, can allow companies to exploit the value of big data in a dramatically improved manner. There is a profound philosophical difference between giving users wide access to data and only place restrictions where needed as opposed to starting with a very limited access and adding to it if and only if there is a compelling business need (somewhat analogous to difference between continental and English common law systems).

Has this philosophy of making data as widely available as possible taken root in Australia? Based on the number of questions asked during the conference about how to restrict access and monitor behaviors I would say we still have a long way to go. Of course there is a need to balance the free access to data with the need for appropriate restrictions and Alex outlined eBay's approach to implementing those restrictions including: permissions, automated monitoring, automated retirement of cold data-marts, productionisation of hot data-marts and the need to pass an exam to get access to Teradata. But it is informative that most questions were about how to manage restrictions rather than on the benefits of open access to data.

Another interesting use of date eBay is the meta-analysis of the queries being submitted. Alex described a program they have to use python to analyze the queries being submitted to Teradata, Singularity and Hadoop. The aim is to identify sub-optimal queries, but also to identify commonly requested information and to develop more efficient ways to deliver this to users. This is an example of a growing trend of data generating data – the data in the warehouse for example indirectly generates or causes users to write queries which themselves then become data which can be analyzed.

Ross Farrelly is the Chief Scientist for Teradata Date ANZ who is responsible for data mining, analytics and advanced modeling projects using the Aster Teradata platform. He is a six sigma black belt and has had many years of experience in a variety of statistical roles.


How to super-power your efforts: learn from Heroes of Analytics

In Hollywood, summer means blockbuster movie ... and with blockbusters like explosions, Car Crash, daredevil stunts and of course superheroes. Summer 2013 is no exception, with titles like the man of steel and The Lone Ranger, the movie audience is for lots of action-packed adventures in the coming months.

Of course, not everyone needs to go to the theater to get their fill of superhero Action. Some of us are fortunate to work in offices where every day there shoulder to shoulder with the superhero – super heroes of data analysis, that is.

These heroes transform large amounts of data from an insurmountable challenge business to actionable, acquired and processed for the revenues of units. As described on the website of heroes of Teradata and SAS Analytics:

"One by one, they emerge from the darkness, snatching the value from the clutches of ambiguity, exposing fraud in a sea of chaos and innovate in the face of statistical impossibility ..."

As you can tell, here at Teradata, we had some fun with the metaphor of the superhero altogether. But that's not to diminish, in any way, the true that Analytics professionals today are committed every day trailblazing work (and a lot of nights, work too).

For example, meet Megavox – otherwise known as Frank Caputo.

As a member of the marketing team at Medibank, Frank used solutions in the Teradata database and SAS to improve the speed of creating campaign from hours to minutes, optimize marketing investments, minimize dependence on external suppliers and more – that led to enormous improvements in the effectiveness of the campaign, together with significant reductions in costs and the substantial increase in revenue.

There is an analytic superhero at your company? Maybe you are one yourself? If so, please consider applying to our Analytics Heroes program. Is your chance to receive hero's fame and respect Analytics, get your own action figure and also be immortalized by a famous comic artist.

Talking about fame and respect, also we are now accepting nominations for awards 2013 Teradata Epic, which recognize our customers and partners for their leadership in implementing data solutions and data analysis.

If your team has created a solution based on a Teradata platform that delivered the bottom-line business value to your company, this is an opportunity to win the recognition they deserve. The application deadline is July 19, 2013, and award winners will be announced and celebrated at the awards ceremony of Teradata EPIC held during Teradata Partners Conference, User Group Expo & this October in Dallas.

How does the superpower google analytics? Apply to Analytics heroes Teradata program and/or an epic Prize for sharing your story ... and who knows, maybe you'll be on your way to saving the world, a campaign based on data at a time.

-Darryl