Apparently there is a word for the system that I’ve been constructing out of various software and hardware: it’s called a “toolchain.” Check out the use of toolchains in ethnography here and here. I’m glad to see that there is another researcher for whom ConnectedText works as a replacement for a mainstream QDA software (in this case Atlas.ti).
I have described the chart below that depicts my use of ConnectedText (CT) for qualitative data analysis as representing a “conceptual model” and “a process flow.” But these terms don’t quite get the idea across that in fact CT had allowed me to construct my very own data analysis machine.
Once you have the basic structure and the logic of this system set up in CT, it works almost like a “sausage machine” with some filters put in. All you need to do is start pumping your empirical data in at Step 1, and as long as you follow the procedure and apply your theoretical filters during the abstraction process, the machine guides you through the production of some “truths,” i.e. your qualitative findings, the answers to your research question.
This brings me back to my earlier points (here and here) about why I prefer to do my qualitative analysis in CT, rather than in Atlas.ti or NVivo. There is no question that those other two dedicated CAQDAS software have more data analysis features and capabilities than CT. However, CT trumps them in one regard hands down: rather than just allowing you to analyse your data, it in effect allows you to create and operate your very own research tools, such as my “idea-sausage machine” below.
Check out my tutorial here, if you are interested in creating your own research tool.
This is my first post in what hopefully will become a series of posts on how I use ConnectedText (CT) for qualitative data analysis, as part of my Ph.D. project. You may ask: “Why bother using CT, a personal wiki, for qualitative data analysis when there are such long-standing dedicated CAQDAS tools on the market such as Atlas.ti and NVivo?” That is a legitimate question. Some people may indeed find that Atlas.ti or NVivo works better for them. However, based on my personal experience as a PhD student, I found that neither Atlas.ti or NVivo was quite right for me. Finding CT was a revelation, as it allowed me to overcome some problems (more about those later) that I couldn’t solve with NVivo, my CAQDAS of choice prior to CT.
How to select a CAQDAS for your study? Try several if you can. Test them on a pilot project. Take workshops on them. Although here I will be recommending CT as CAQDAS for particular qualitative analysis jobs, I still suggest you use either Atlas.ti or NVivo extensively for a prolonged period of time to learn about how mainstream CAQDAS work.
What was my experience with CAQDAS? I took a couple of workshops in Atlas.ti, got a copy and trialled it repeatedly over the years. As far as I know, Atlas.ti was originally developed to implement grounded theory, and as I wasn’t completely convinced by its implicit research philosophy (which still seemed a bit positivist to me, although otherwise I’m all in favour of grounded theory’s bottom-up approach), I found some of the lingo, the design and the interface off-putting. I discovered that I was a visual learner and found that Atlas.ti wasn’t accommodating my type of learner. I found it difficult to visualise the conceptual linkages between Atlas.ti’s various commands and floating windows.
NVivo 8, on the other hand, immediately appealed to me the first time I laid my eyes on it. I could see how its different tools related to each other. At its heart there was a hierarchical folder structure that worked like many other Windows-based software. It also seemed more neutral from a research philosophy perspective. In the end I spent 6 months analysing my qualitative data in NVivo. I used it to code about half of my data (interviews, participant observations, and a variety of collected documents in a range of media and file formats). I enjoyed doing the coding with it. [The screenshots below are of the example database that comes with NVivo.]
Why did I abandon NVivo? The main reason had to do with the limitations of NVivo’s interface, or rather the way its tools are organised and the way this organisation forces you to follow a particular analytical logic. The underlying issue is the hierarchical organisation of folders and tools, which, ironically, was the reason I chose NVivo over Atlas.ti in the first place. The problem was that different elements of the work get separated into different folders, which can’t be viewed at the same time, thus breaking up the workflow and the analytical reasoning.
After all my coding I ended up with a hierarchical forest of codes (a lot more complex one than the example in the above image) and it was difficult to see how to bring them all together again in order to synthesise them into findings and a coherent answer to my research question. I was stuck and confused, as I also lost my overview of all the data that I had reviewed and stored in there. (In fact, in my desperation I took screenshots of my NVivo codes and pasted them into PowerPoint, and continued the rest of the work outside of NVivo. It helped to cut a Gordian knot but I still ended up needing another CAQDAS solution for the rest of my data).
I also realised that NVivo is a wiki forced into a hierarchical straight-jacket. You are essentially creating links between different documents and then these links are presented to you in a hierarchical organisation. This rekindled my interest in personal wikis again. Since the early days of my PhD I wished that there was a way to create a dashboard for my dissertation project, from which all my files (both data and my analysis and writings) would be linked in the manner of an intranet. I have experimented with Planz, WhizFolders and a range of desktop wikis but none of them were quite right. In fact I also checked out ConnectedText a few times over the years but I never managed to make sense of it. However, my problems with NVivo’s hierarchical structure convinced me to have another go with the wiki format. Fortuitously, there were a series of helpful posts on Outliner Software and on his blog by Steve Zeoli, which finally led me to an epiphany with CT. But more about that in my next post.
There were some other reasons too why I was happy to say good-bye to NVivo (and the same is true for Atlas.ti). One was the enormous price tag and their licensing regime. Although I got a very cheap, subsidised license from my university, it needs to be renewed annually, and after I graduate I may need to shell out some serious money if I want to have continued access to the data in the form it’s organised in NVivo. To me it seems that NVivo and Atlas.ti to some extent cornered the market and are taking advantage of the fact that PhD students are a captive audience, strongly influenced by the preferences of their supervisors, peers and academic tribes. This business attitude I find very unappealing.
Finally, there are the resource requirements of running a software like NVivo. When I installed my first NVivo copy onto my previous computer, a laptop, I wasn’t even able to launch it. It just simply wouldn’t budge and froze my entire machine. I was forced to buy a top-of-the-range PC in order to be able to run it, and even on the new machine it was sluggish. Waiting for files and windows to open for several seconds eventually adds up. Also, as far as I know, all the work in NVivo is saved into a single proprietary file. It is just asking for a lot of faith to put several years worth of data and effort into a single massive file and hope that it will work for 100% of the time forever, that it will never get corrupted etc., etc. I’m not a techie but it seems to me there are some risks associated with that approach.
Still, I’m glad that I had a proper go with NVivo. It was only by spending many months toiling with it that I had learned how CAQDAS work and what the features are that I like and the ones that I would like to have. I was only able to adopt CT as my new CAQDAS tool because I was able to model the processes that I learnt about – or wasn’t allowed to do – in NVivo.
In my next two posts I will introduce the key qualitative data analysis features of CT and I will also provide some tips on how to get started with it, especially if you had never worked with a wiki before.