Sunday, 9 November 2014

Big Data, Small Data and Meaning

This post was originally written as the text for a talk I gave at a British Library Lab's event in London in early November 2014. In the nature of these things, the rhythms of speech and the verbal ticks of public speaking remain in the prose. It has been lightly edited, but the point remains the same.  

In recent months there has been a lot of talk about big stuff. Between 'Big Data' and calls for a return to ‘Longue durée’ history writing, lots of people seem to be trying to carve out their own small bit of 'big data'. This post represents a reflection on what feels to me to be an important emerging strategy for information interrogation driven by the arrival of 'big data' (a 'macroscope'); and a tentative step beyond that, to ask what is lost by focusing exclusively on the very large. 

And the place I need to start is with the emergence of what feels to me like an increasingly commonplace label – a ‘macroscope’ - for a core aspiration of a lot of people working in the Digital Humanities. 

As far as I can tell, the term ‘macroscope’ was coined in 1969 by Piers Jacob, and used as the title of his science fiction/fantasy work of the same year – in which the ‘macroscope’, a large crystal, able to focus on any location in space-time with profound clarity, is used to produce something like a telescope of infinite resolution. In other words, a way of viewing the world that encompasses both the minuscule, and the massive. The term was also taken up by Joel de Rosnay and deployed as the title of a provocative book on systems analysis first published in 1979. The label has also had a long and undistinguished afterlife as the trademark for a suite of project management tools – a ‘methodology suite’ - supported by the Fujistu Corporation. 

But I think the starting point for interest in the possibility of creating a ‘macroscope’ for the Digital Humanities, comes out of computer science, and the work of Katy Börner from around 2011.
Her designs and advocacy for the development of a ‘Plug and Play Macroscope’, seems to have popularised the idea to a wider group of Digital Humanists and developers. To quote Börner

'Macroscopes provide a "vision of the whole," helping us "synthesize" the related elements and detect patterns, trends, and outliers while granting access to myriad details. Rather than make things larger or smaller, macroscopes let us observe what is at once too great, slow, or complex for the human eye and mind to notice and comprehend.' (Katy Börner, ‘Plug-and-Play Macroscopes’, Communications of the ACM, Vol. 54 No. 3, Pages 60-6910.1145/1897852.1897871)

In other words, for Börner, a macroscope is a visualisation tool that allows a single data point, to be both visualised at scale in the context of a billion other data points, and drilled down to its smallest compass. This was not a vision or project initially developed in the humanities. Instead it was a response to the conundrums of ‘Big Data’ in both STEM academic disciplines, and the wider commercial world of information management. But more recently, a series of ‘macroscope’ projects have begun to emerge from within the humanities, tied to their own intellectual agendas, and subtly recreating the idea with a series of distinct emphases. 

Perhaps the project most heavily promoted recently, is Paper Machines, created by Jo Guldi and Chris Johnson-Robertson – and the MetLab at Harvard. This forms a series of visualisation tools, built to work with Zotero, and ideally allowing the user to both curate a large scale collection of works, and explore its characteristics through space, time and word usage. In other words, it is designed to allow you to build your own Google Books, and explore. There are problems with Paper Machines, and most people I know have struggled to make it work consistently. But it rather nicely builds on the back of functionality made available through Zotero, and effectively illustrates what might be described as a tool for ‘distant reading’ that encompasses elements of a ‘macroscope’. 

What is most interesting about it, however, is the use its creators make of it in seeking to shift a wider humanist discussion from one scale of enquiry to another. Last month, to great fanfare, CUP published Jo Guldi and David Armitage’s History Manifesto, which argues that once armed with a ‘macroscrope’ – Paper Machines in their estimation historians should pursue an analysis of how ‘big data’ might be used to re-negotiate the role of the historian – and the humanities more generally. Basically, what Guldi and Armitage are calling for through both the Manifesto and through Paper Machines, is the re-invention of ‘Longue durée’ history – telling ever larger narratives about grand sweeps of historical change, encompassing millennia of human experience. And to do this in pursuit of taking on the mantle of a public intellectual, able to speak with greater authority to ‘power’. 

In the process they explicitly denigrate notions of ‘micro-history’ as essentially irrelevant. At one and the same time, they seem to me to celebrate the possibility of creating a ‘macroscope’, while abjuring half its purpose. What we see in this particular version of a ‘macroscope’ is a tool that privileges only one setting on the scale between a single data point, and the sum of the largest data set we can encompass. In other words, by seeking the biggest of big stories, it is missing the rest. 

Perhaps the other most eloquent advocate for a ‘macroscope’ at the minute is Scott Weingart. With Shawn Graham and Ian Milligan, he is writing a collective online ‘book’ entitled, Big Digital History: Exploring Big Data through a Historian’s Macroscope. The book is a nice run through of digital humanist tools, but the important text from my perspective is a blog post Weingart published on the 14 September 2014. The post was called: The moral role of DH in a data-driven world; and in it, Weingart advocates a very specific vision of a ‘macroscope’, in which the largest scale of reference and view is made intelligible through the application of a formal version of network analysis. 

Weingart is a convincing advocate for network analysis, performed in light of some serious and sophisticated automated measures of distance and direction. And his work is a long way ahead of much of the naïve and unconvincing use of network visualisations current in large parts of the Digital Humanities. Weingart also makes a powerful case for where a limited number of DH tools – primarily network analysis and topic modelling - could be deployed in re-engaging the ‘humanities’ with a broader social discussion. 

Again, like Guldi and Armitage, Weingart seeks in 'Big Data' a means through which the Humanities can ‘speak to power’. As with the work of Armitage and Guldi, the pressing need to turn Digital Humanities to political account appears to motivate a search for large scale results that can be deployed in competition with the powerful voices of a positivist STEM tradition. My sense is that Weingart, Armitage and Guldi are all essentially scanning the current range of digital tools, and selectively emphasising those that feel familiar from the more ‘Social Science’ end of the Humanities. And that having located a few of them, they are advocating we adopt them in order to secure our place at the table. 

In other words, there is a cultural/political negotiation going on in these developments and projects that is driven by a laudable desire for ‘relevance’, but which effectively moves the Humanities in the direction of a more formal variety of Social Science. 

Others still, are arguably doing some of the same work, but using a different language, or at the least seeking a different kind of audience. Jerome Dobson, for example, has recently begun to describe the use of Geographical Information Systems (GIS) in historical geography, as a form of ‘macroscope’. This usage doesn’t come freighted with the same political claims as are current in Digital Humanities, but seem to me an entirely reasonable way of highlighting some of the global ambitions – and sensitivity to scale - that are inherent in GIS. The notion - perhaps fostered most fully by Google Earth - that you can both see the world in its entirety, as well as zoom in to the smallest detail, seems at one with a data driven ‘macroscope’. But, again, the scale most geographers want to work with is large – patterns derived from billions of data points. And again, the siren call of GIS, tends to pull humanist enquiry towards a specific form of social science. 

And finally, we might also think of the approach exemplified in the work of Ben Schmidt as another example of a ‘macroscope’ approach – particularly his ‘prochronism’ projects. These take individual words in modern cinema and television scripts that purport to represent past events – things like Downton Abbey and Mad Men - and compares them to every word published in the year they are meant to represent. 

Building on Google Books and Google Ngrams, Schmidt is effectively mixing scales of analysis at the extremes of ‘big data’, on the one hand – all words published in a single year – and small data, on the other. Of all the examples mentioned so far, it is only Schmidt who is actually using the functionality of a ‘macroscope’ effectively, making it all the more ironic that he doesn’t adopt the term. 

And almost uniquely in the Digital Humanities – a field equally remarkable for its febrile excitement, and lack of demonstrable results – Schmidt’s results have been starkly revealing. My favourite example, is his analysis of the scripts of Mad Men, which illustrates that early episodes referencing the 1950s, overuse language associated with the ‘performance’ of masculinity – words that reflect ‘behaviour’. And that later episodes, located in the 1970s, overuse words reflecting the internalised emotional experience of masculinity. For me this revealed beautifully the larger narrative arc of the programme in a way that had not been obvious prior to his work. Schmidt has little of the wider agenda to influence policy and politics evident in that of Armitage, Guldi and Weingart, but ironically, it is his work that is having some of the greatest extra-academic impact, via the anxiety it has created in the script writers of the shows he analyses. 

All of which is simply to say that playing with and implementing ideas around a ’macroscope’ is quite popular at the moment. And a direction of travel which, with caveats, I wholly support. But it also leaves me in something of a conundrum. 

Each of these initiatives, with the possible exception of Schmidt’s work, seems to locate themselves somewhere other than the Humanities I am familiar with. And this seems odd. Issues of scale are central to this. Claiming to be doing ‘big history’ sounds exciting; while claiming that more formal ‘network analysis’, will answer the questions of a humanist enquiry, appears to create a bridge between disciplines – allowing Humanists and more data driven parts of the Social Sciences to share a methodology and a conversation. But with the exception of Schmidt’s work, these endeavours seem to be privileging particular types of analysis – Social Science types of analysis – over more traditionally Humanist ones. 

In some ways, this is fine. I have discovered to my own benefit, that working with ‘Big Data’ at scale and sharing methodologies with other disciplines is both hugely productive, and hugely fun. To the extent that ‘big stories’ and new methodologies provide the justification for collaborating with researchers from a variety of disciplines – statisticians, mathematicians and computer scientists – they are wholly positive, and a simple ‘good thing’. 

And yet… I find myself feeling that in the rush to define how we use a ‘macroscope’, we are losing touch with what humanist scholars have traditionally done best. 

I end up is feeling that in the rush to new tools and ‘Big Data’ Humanist scholars are forgetting what they spent much of the second half of the twentieth century discovering – that language and art, cultural construction, human experience, and representation are hugely complex – but can be made to yield remarkable insight through close analysis. In other words, while the Humanities and ‘Big Data’ absolutely need to have a conversation; the subject of that conversation needs to change, and to encompass close reading and small data. 

The Stanford Humanities Centre defines the ‘Humanities’ as: 

'…the study of how people process and document the human experience. Since humans have been able, we have used philosophy, literature, religion, art, music, history and language to understand and record our world.'

Which makes the Humanities sound like the most un-exciting, ill-defined, unsalted, intellectual porridge ever. And yet, when I think about the scholarly works that have shaped my life, there is none of this intellectual cream of wheat. 

Instead, there are a series of brilliant analyses that build from beautifully observed detail at the smallest of scales. I look back to the British Marxist tradition in history – to Raphael Samuel and Edward Thompson – and what I see are closely described lives, built from fragments and details, made emotionally compelling by being woven into ever more precise fabrics of explanation. 

A gesture, a phrase, a word, an aching back, a distinctive tattoo. 'My dearest …. Remember when…' 

The real power of work in this tradition, lay in its ability to deploy emotive and powerful detail in the context of the largest of political and economic stories. And the political project that underpinned it, was not to ‘speak to power’, but to mobilise the powerless, and democratise identity and belonging. With Thompson’s liquid prose, a single poor, long dead framework knitter affected more change than any amount of more formal economic history. 

Or I think of the work of Pierre Bourdieau, Arlette Farge and de Certeau, and the ways in which they again use the tiny fragments of everyday life - the narratives of everyday experience - to build a compelling framework illustrating the currents and sub-structures of power. 

Or I think of Michel Foucault, who was able to turn on its head every phrase and telling line – to let us see patterns in language – discourses – that controlled our thoughts. Foucault profoundly challenged us to escape the limits of the very technologies of communication and analysis we used; and to see in every language act, every phrase and word, something of politics. 

By locating the use of a ‘macroscope’ at the larger scale, seeking the Longue durée, and the ear of policy makers, recent calls for how we choose to deploy the tools of the Digital Humanities appear to deny the most powerful politics of the Humanities. If today we have a public dialogue that gives voice to the traditionally excluded and silenced – women, and minorities of ethnicity, belief and dis/ability – it is in no small part because we now have beautiful histories of small things. In other words, it has been the close and narrow reading of human experience that has done most to give voice to people excluded from ‘power’ by class, gender and race. 

Besides simply reflecting a powerful form of analysis, when I return to those older scholarly projects I also see the yearning for a kind of ‘macroscope’. Each of these writers strive to locate the minuscule in the massive; the smallest gesture in its largest context; to encompass the peculiar and eccentric in the average and statistically significant. 

What I don’t see in modern macroscope projects is a recognition of the power of the particular; or as William Blake would have it: 

To see a World in a grain of sand, 
And a Heaven in a wild flower...
                               Auguries of Innocence (1803, 1863).

Current iterations of the idea of a macroscope, with all their flashy, shock and awe visualisations, probably score over these older technologies of knowing in their sure grasp of data at scale, but in the process they seem to lose the ability to refocus effectively. 

For all the promise of balancing large and small scales, the smaller and particular seem to have been ignored. Ever since the Apollo 17 sent back its pictures of earth as a distant blue marble, our urge towards the all-inclusive, global and universal has been irresistible. I guess my worry is that in the process we are losing the ability to use fine detail in the ways that make the work of Thompson and Bourdieau, Foucault and Samuel, so compelling. 

So, by way of wending towards some kind of inconclusive conclusion. I just want to suggest that if we are to use the tools of 'Big Data' to capture a global image, it needs to be balanced with the view from the other end of the macroscope (along with every point in between). 

In part this is just about having self-confidence as humanist scholars, and ironically serving a specific role in the process of knowing, that people in STEM are frequently not very good at. 

Several recent projects I was privileged to participate in, involved some hugely fun work with mathematicians and information scientists exploring the changing linguistic patterns found in the Old Bailey trials – all 127 million words worth. And after a couple of years of working closely with a bunch of brilliant people, what I gradually realised was that while mathematicians do a lot of ‘close reading’ – of formulae and algorithms - like most scientists, they are less interested than I am in the close reading of a single datum. In STEM cleaning data is a chore. Geneticists don’t read the human genome base by base; and our knowledge of the Higgs Boson is built on a probability only discovered after a million rolls of the dice, with no one really looking too carefully at any single one. 

In many respects ‘big data’ actually reinforces this tendency, as the assumption is that the ‘signal’ will come through, despite the noise created by outliers and weirdness. In other words, ‘Big Data’ supposedly lets you get away with dirty data.  In contrast, humanists do read the data; and do so with a sharp eye for its individual rhythms and peculiarities – its weirdness. 

In the rush towards 'Big Data' – the Longue durée, and automated network analysis; towards a vision of Humanist scholarship in which Bayesian probability is as significant as biblical allusion, the most urgent need seems to me to be to find the tools that allow us to do the job of close reading of all the small data that goes to make the bigger variety. This is not a call to return to some mythical golden age of the lone scholar in the dusty archive – going gradually blind in pursuit of the banal. This is not about ignoring the digital; but a call to remember the importance of the digital tools that allow us to think small; at the same time as we are generating tools to imagine big. 

In relation to text, you would think this is easy enough. Easy enough to, like Ben Schmidt, test each word against its chronological bed-fellows; or measure its distance from an average for its genre. When I am reading a freighted phrase from the 1770s, like ‘pursuit of happiness’, I want to know that till then, ‘happiness’ was almost exclusively used in a religious context – ‘Eternal Happiness’ - and that its use in a secular setting would have caught in a reader’s mind as odd and different - new. We should be able to mark the moment when Thomas Jefferson allowed a single word to escape from one ‘discourse’ and enter another – to read that word in all its individual complexity, while seeing it both close and far. 

I know of no work designed to define the content of a ‘discourse’, and map it back in to inherited texts. I know of no projects designed with this notion in mind. And if you want a take home a message from this post, it is a simple call for ‘radical contextualisation’. 

 To do justice to the aspirations of a macroscope, and to use it to perform the Humanities effectively – and politically – we need to be able to contextualise every single word in a representation of every word, ever. Every gesture contextualised in the collective record all gestures; and every brushstroke, in the collective knowledge of every painting. 

Where is the tool and data set that lets you see how a single stroll along a boulevard, compares to all the other weary footsteps? And compares it in turn to all the text created along that path, or connected to that foot through nerve and brain and consciousness. Where is the tool and project that contextualises our experience of each point on the map, every brush stroke, and museum object? 

This is not just about doing the same old thing – of trying to outdo Thompson as a stylist, or Foucault for sheer cultural shock. My favourite tiny fragment of meaning – the kind of thing I want to find a context for - comes out of Linguistics. It is new to me, and seems a powerful thing: Voice Onset Timing – that breathy gap between when you open your mouth to speak, and when the first sibilant emerges. This apparently changes depending on who are speaking to – a figure of authority, a friend, a lover. It is as if the gestures of everyday life can also be seen as encoded in the lightest breathe. Different VOTs mark racial and gender interactions, insider talk, and public talk.

In other words, in just a couple of milliseconds of empty space there is a new form of close reading that demands radical contextualisation (I am grateful to Norma Mendoza-Denton for introducing me to VOT). And the same kind of thing could be extended to almost anything. The mark left by a chisel is certainly, by definition, unique, but it is also freighted with information about the tool that made it, the wood and the forest from which it emerged; the stroke, the weather on the day, and the craftsman. 

One of the great ironies of the moment is that in the rush to big data – in the rush to encompass the largest scale, we are excluding 99% of the data that is there. And if we are going to build a few macroscopes, I just want to suggest that, along with the blue marble views, we keep hold of the smallest details. And if we do so, looking ever more closely at the data itself – remembering that close reading can be hugely powerful - Humanists will have something to bring to the table, something they do better than any other discipline. They can provide a world of ‘small data’ and more importantly, of meaning, to balance out the global and the universal – to provide counterpoint in the particular, to the ever more banal world of the average.


  1. You wouldn't ask Hans Holbein to lead your troops into battle. You'd give him a paintbrush, and get the hell out of the way.

  2. This is an eloquently written post, and I think the views it's expressing will be eagerly received by historians and other humanists. I also love the metaphor of the macroscope, and appreciate discovering Katy Börner as an important source.

    But to my way of thinking there's a deep internal contradiction in prevailing humanistic opinions about macroscopic analysis, and I have to admit that I feel traces of that contradiction here as well. On the one hand, we admit that there are still only a handful of people doing this new kind of work. On the other hand, we worry that it's going to rule out, or exclude, or steal all the oxygen from, more traditional scales of humanistic analysis. Which would be quite a trick for three or four junior scholars to perform!

    I can imagine two ways of justifying this contradiction. On the one hand, we could say, "it's a slippery slope; if humanists start down this path we will be unable to stop." On the other hand, we could say "I know macroscopic research is still a tiny subset of humanistic scholarship, and perhaps it may remain so. But it's important for close reading to play an equal role even in this tiny subset of works. We can't afford to have even a *small* number of articles where macroscopic analysis is privileged over close reading, because in the humanities (although not the social sciences) it would be a fundamental error."

    But I may be reasoning ungenerously here; I'm curious how you would respond to this puzzle. How can the works of a small number of scholars pose a methodological threat to the humanities as such?

  3. Thanks, Tim, for a really thoughtful piece as always. I think you and I agree about everything, or most everything, here, but I think maybe I haven't been as clear as I'd like to be in earlier pieces. You suggest macroscopic thinking stears us way too close to social science, and too far away from the small scale, which is exactly the opposite of the message I'm hoping to get across.

    There are two claims being made on my, Guldi's, and Armitage's behalf, and I'd like to tease those apart before addressing them. The first (1) is that we advocate focusing on the large at the expense of the small, and the second (2) is that adopting such methods will "secure our place at the table", making us relevant again. The result of our actions, this post suggests, will lead us away from the humanities and towards the social sciences.

    While I can't speak for Armitage or Guldi, I think these claims don't quite match my own arguments. I think I may be at fault for the disparity, but I'll address that later.

    The idea that (1) the large should be researched at the expense of the small is something I argue against as often as I can. In the "Moral Role" post you linked to, I argue instead that the digital humanist can inject the small back into the large, offering a much-needed humanistic eye in an otherwise social scientific playground.

    As for discussions of (2) relevance and place at the table, while that is something I believe macroscopic thinking is currently helping with, I don't suggest it's a goal of DH as much as an accidental outcome. The argument is not that DH has a moral imperative to kick the humanities, kicking and screaming, into relevancy; the argument is that there are a growing number of ethical concerns in this world that people with DH training are well-suited to address, and those people should address them. Whether this should be a goal of the humanities in general, or DH in particular, is besides the point.

    You later go on to suggest the macroscopic stance tends to privilege social scientific analysis over humanities varieties. I don't think this is the case, but I believe those of us who advocate such a stance are guilty of making that difficult to see. In blog posts and tools and twitter and conferences/journals to other DHers, where most of this macroscopic groundwork gets done, we make the mistake of excitedly talking about methods and tools without addressing the big picture.

    It seems like all we talk about is social scientific methods. But there is (or should be) a difference in what we present to our fellow DHers vs. what we publish for popular or humanistic consumption - the actual results of our research rather than the discussion of our methodologies. I think most in the "macroscope" camp, myself included, would argue that good macroscopic research should be integrative, taking advantage of any methods useful to the task at hand - usually a combination of the quantitative and the qualitative, the humanistic and the social scientific.

    Unfortunately, I fear this point doesn't come across in our in-group writing, which is something we ought to address.

    "Where is the tool and data set that lets you see how a single stroll along a boulevard, compares to all the other weary footsteps?" you ask. This is the problem we're all trying to figure out, I think.

  4. Dear Ted and Scott, Thanks for these comments - both of which are well made and entirely right. For myself, all these different varieties of history writing and the humanities are valuable - both those directed at policy and the small scale study emphasising the shock of the old. I have bridled a bit at the enthusiasms of some scholars, but it would be wrong to advocate any single type of analysis to the exclusion of the others.

    The kind of keynote talk that this post was based on tends to demand overstatement, and I think you both effectively point up where I have overstated! Having said this, I do want to meld the small with the big, and am genuinely struggling to figure out how to write emotionally and poliitcally engaged history using all the tools I can find.

    In part, I suspect that this is as much about genre and writing as it is about tools and DH, or even the politics of academic engagement. The technology makes a macroscope possible, but we are still writing monographs of one sort or another (whether meta or micro). I would love to find the form of writing and presentation that makes full use of all the scales. But, in the absence of that, I think Ted is right to point up the need to generously encourage the lot; and Scott is right that perhaps all I am pointing to is our tendency to talk about tools and methods, without talking enough about humanist analysis.

    In a UK the enthusiasm for the big, and 'big data', feels like a siren call that needs to be resisted.

    My problem is that I really like 'big data' approaches, and think they offer something real and different (an 'open eyed' stare at stuff); but the minute I go down that route, I immediately miss the small, close reading, and the power of detail to evoke emotion and empathy.

    Thanks for the engagement!

  5. Just riffing off of Scott's reply here, in relation to the question of connecting scales of analysis:

    I actually don't think it's at all hard to pause macroscopic analysis for a close-up view of particulars. No special tool is required.

    But sometimes I pause for close readings, and sometimes I don't. The reason has nothing to do with technology; it's a question of genre, subject, and available space. I use "close-ups" if I think they're important for a particular topic, or if they make the writing livelier. In a book-length argument I certainly do a lot of it. But in shorter genres, lately, I often don't; I often feel that I can make a livelier or clearer argument by staying at the macroscopic scale.

    And I guess I feel not even the tiniest twinge of remorse about that. It's possible that it does mean I'm working in a more social-scientific mode in some pieces. Conversely, Thomas Piketty's latest book-length work strays from economics, and starts to poach on the turf of the humanities. My view: it's all good.

  6. Of course, the idea that becoming social science will make us relevant, and listened to, would have a conference of actual social scientists rolling, and weeping, in the aisles...

  7. Thanks for a great post urging us to think about the tensions inherent in digital humanities practices. I'm also glad to see further discussion unfolding in the comments.

    As I read the post, I found it resonating with some issues I've been thinking about for a while, about the necessity of moving between the micro- and macroscales, and about what it means to do this type of work from a humanist/humanities perspective. A while back, I wrote some reflections on similar ideas, from a different perspective, in relation to some of my own work here. Much of my own work, in fact, came out of thinking about how computational methods of analyzing big data could also be harnessed for analyzing a small corpus of data (my own few dozen texts). I'm still working through some of those questions, so this post is useful for focusing some of my thoughts for me.

    There's certainly a lot more to be said about these possibilities and tensions, and for reconciling the micro and macro scales. I look forward to seeing the conversation continue to unfold.

  8. Wiztech Automation is a Chennai based one-stop Training Centre/Institute for the Students Looking for Practically Oriented Training in Industrial Automation PLC, SCADA, DCS, HMI, VFD,VLSI, Embedded, and others – IT Software, Web Designing and SEO.

    PLC Training in Chennai
    Embedded Training in Chennai
    VLSI Training in Chennai
    DCS Training in Chennai
    IT Training Institutes in Chennai
    Web Designing Training in Chennai

  9. Big data is hard to analyze. These are large volume of data that is so large and complex.
    Big Data

  10. So the main ideas have already been cited, it will be good for the students to carry out more guidelines before going to decide anything. terms of service generator

  11. The term often refers simply to the use of predictive analytics, user behaviour analytics, or certain other advanced data analytics methods that extract value from data, and seldom to a particular size of data set. Thanks a lot admin to sharing with us. Also Learn BigData from the best BigData Online Training in your locality at

  12. First of all i am saying that i like your post very much.I am really impressed by the way in which you presented the content and also the structure of the post. Hope you can gave us more posts like this and i really appreciate your hardwork.

    Hug Day 217
    Happy New Year 2018 Poems
    broken heart dp for whatsapp
    Good Night SMS
    Happy Birthday Brother Wishes
    Happy New Year 2017 Quotes
    Valentine Week List 2017

  13. thank you sharing valuable information with us.we are very glad to leave a comment here.thank you for offering such nice of the recommanded

    Hadoop and Bigdata online training
    Hadoop training
    Hadoop and Bigdata training

  14. This is a very good article very useful thank you for sharing. Also know about Big Data and Hadoop Online Training

  15. Look at the product pipeline, look at the fantastic financial results we've had for the last five years. You only get that kind of performance on the innovation side, on the financial side, if you're really listening and reacting to the best ideas of the people we more:-Shenglin Financial

  16. The increase of mobile phone usage in emerging markets makes mobile money a safe, low cost initiative for the financial sector. It is an easier way to transfer money to family and friends, money is sent, and payments and withdrawals can be made without ever going to a physical bank or payment center.To know more at:Shenglin Financial

  17. Thank you for your clear explanation between the big data and small data keep share the valuable thoughts like this

    Graphic designer jobs in bangalore | Web designer jobs in bangalore

  18. Just found your post by searching on the Google, I am Impressed and Learned Lot of new thing from your post. I am new to blogging and always try to learn new skill as I believe that blogging is the full time job for learning new things day by day.
    "Emergers Technologies"

  19. Thanks for sharing Valuable information about Bigdata. This post is really helped me a lot. Keep sharing........... If it possible share some more tutorials???????????

  20. This comment has been removed by the author.

  21. This blog is the general information for the feature. You got a good work for these blog.We have a developing our creative content of this mind.Thank you for this blog. This for very interesting and useful.
    Analytics Training in Chennai

  22. It’s a very good artical and good creation and good information thanks for sharing
    Social media seminar

  23. Thank you for taking the time and sharing this information with us. It was indeed very helpful and insightful while being straight forward and to the point. | |

  24. Excellent blog,thank you for the useful information....
    Hadoop Training institute in chennai

  25. Good blogger and thanks for sharing information

  26. I am really impressed the way you have written the blog. Hope we are eagerly waiting for such post from your side. HATS OFF for the valuable information shared!
    Hadoop Training in Chennai
    Hadoop Training Chennai
    Big Data Training in Chennai

  27. This is a great post. I like this topic.This site has lots of advantage.I found many interesting things from this site. It helps me in many ways.Thanks for posting this again.

    Hadoop Training in Chennai

    Base SAS Training in Chennai

    MSBI Training in Chennai

  28. keep sharing this type of useful information. This is best blog for Big Data and Hadoop Developers.

    Thank you
    Big Data Training in Hyderabad

  29. Hi,
    Thanks for sharing such a wonderful information on
    "Big Data, Small Data and Meaning"
    Keep sharing such a wonderful articles with us.
    Big Data Analytics Training In Hyderabad
    Big Data Analytics Course In Hyderabad

  30. Hi,
    A very Impressive and detailed post about the Big Data, Small Data meaning.Thanks for sharing such good and useful information to Big Data Hadoop users and developers.
    Hadoop Training in Hyderabad

  31. I am really happy to say it’s an interesting post to read . I learn new information from your article , you are doing a great job . Keep it up

    Data Science Training in Hyderabad
    Hadoop Training in Hyderabad

  32. Hello sir, your web site is lovingly serviced and saved as much as date. So it should be, thanks for sharing this with us.
    i also found some helpful sites like yours.

    irctc login



    Indian Railway

    Train enquiry

    PNR Status

    Pnr enquiry

    running status

    spot your train
    live train status
    Train between stations
    book train tickets

  33. Nice blog .This has give a clear picture on origin of Big Data big data hadoop online training

  34. Hello sir, your web site is lovingly serviced and saved as much as date. So it should be, thanks for sharing this with us.
    i also found some helpful sites like yours.

    Govt jobs

    free job,
    railway recruitment
    ibps jobs
    job alerts

    bank jobs

    ssc jobs

    railway recruitment 2017

    latest jobs

    upcoming jobs

    army jobs

  35. Your good knowledge and kindness in playing with all the pieces were very useful. I don’t know what I would have done if I had not encountered such a step like this. Big Data Training Institute in Chennai

  36. Nice blog. Really helpful for learning Big Data Training and keep update on some more tutorials….. I liked your blog.

  37. Thanks for sharing such awesome posts on this blog. i am really happy to read about Easter Day 2018 Images

  38. Well written article. Get rid of the traditional ways of marketing and buy a customized email list from us. One real-time data can change your entire business for good. Parana Impact provides you the Email List that you are exactly looking for.
    We have worked with many organizations of all the sizes. Our main focus is customer satisfaction. Big Data Users Email List