The purpose of this report is to explore the future of Information Technology (IT) and its effect on Healthcare Information Systems (HIS), by looking at the present, and future trends in healthcare. Looking at the implementation of HISs, the opportunities and challenges are discussed, and the necessary skill for developing such a system are analysed. Finally, the digital transformation of South London and Maudsley NHS Foundation Trust is reviewed, what was presented at City, University of London by Stephen Docherty Chief Information Office (CIO).
This aim of this report is to explore the future of the IT industry and Artificial Intelligence (AI)and
their effect on
the healthcare industry. First, the terms AI and data mining will be introduced and defined, then the
actual trends in
health care will be discussed in details.
In the literature review, the scientific views of the current opportunities of applying good computer
science practices
will be dissected, as well as the applications of data mining in healthcare. The review will further
investigate what
the main challenges are of this digital transformation. In the discussion, the British trends will be
discussed in
health informatics, and how the legal and ethical issues relate to computer science in healthcare. Finally,
the
required technical and non-technical skills will be analysed, that are required for a computer science
professional in
healthcare.
Artificial Intelligence (AI) has been around since the 1950s, when Alan Turing has written an article about
artificial
intelligence, titled as Computer Machinery and Intelligence (Turing, 1950).
In this article, Turing started by asking the question: “Can machines think?” Since it was difficult to
define
“machine” or “think”, Turing proposed another question: “Can a machine win a game, called ‘Imitation
Game’?” The
Imitation Game (according to Turing) is played by three players: Player A as a man, player B as a woman,
and the
interrogator C. The interrogator can ask questions from the players however, the messages are passed by
notes,
therefore player C cannot see which player is a man or woman. During the game player B’s goal is to
convince the
interrogator that she is a woman, and the goal of A is to trick C. Turing was curious, what would happen if
player A’s
role would be taken by a computer.
Later, the original test was modified, and now several versions exist. For example, in one version, both
player A and B
are trying to convince the interrogator that they are the human. Player C has to decide which player is the
computer.
To pass the test, a computer must trick a human more than 30% of the time during a series of five minutes
keyboard
conversations. First time in the history, this was achieved in 2014 on an event organised by University of
Reading
(University of Reading, 2017).
Part of artificial intelligence is data mining, which can be defined as the process of finding previously
unknown
patterns in large datasets, and then by using these patterns, a predictive model then can be built
(Kincade, 1998).
Data mining is especially powerful on medical data, because usually medical datasets have hundreds of
thousands of data
points.
As it was prior discussed, AI itself has been around for a while, and many of the applied statistical and
mathematical
models were discovered decades ago, however, the field did not have many practical applications, simply
because of the
lack of computer power, and data. However, the processing power and the memory of the computers have
dramatically
improved in the last couple of decades (Moravec, 1998).
Besides the computational power, the available digital data has also been increasing in the past ten years.
In 2014,
the International Data Corporation (IDC) has conducted a study for Dell EMC, and predicted that the amount
of digital
data will double in every two years, and by 2020, the data we create annually will reach 44 trillion Gb
(Zwolenski and
Weatherill, 2014).
To investigate the opportunities for computer science and AI in healthcare, first, one has to consider the
main trends
in health care that support digital transformation.
Moving from a paper-based system to more computer-based systems is common today over the major industries.
In the
medical industry in the past, most of the data was recorded and stored mainly on paper however, the
migration to
paperless approach has already been started (Haux et al., 2002). Increasing the computer system’s
boundaries, has also
been a trend, especially with the appearance of the Internet. While in the past, an information system in
health care
was limited to a department, or to a particular hospital, now new systems are targeting to record whole
cities, regions
or even countries (Linberg, 1968). This expansion of the system boundaries also has an effect on the type
of users of a
system. While earlier, a specific system was designed for the physicians and the administrative workers,
later it was
passed on to be used by nurses (Ball et al., 1994). Today, even patients might have some limited access to
check their
test results, or to consult with doctors. Finally, the way as medical data is used has changed from using
it only to
patient’s care to use it for research or educational purposes (Leiner et al., 2002; Kuhn and Guise,
2001).
As the consequence of these trends, today, the available digital data is more than ever, and as it was
mentioned
before, this is considered to be one of the main requirements for building an effective data mining tool.
Healthcare Information Systems (HIS) refers to the complete information processing and information storing
subsystem of
a hospital (Haux et al., 2004). The medical data of a patient is stored in an Electronic Health Record
(EHR), and
mainly it is either clinical data such as medical test results and notes from consultations, or social
data, like notes
and reports from home visits. With advanced HIS, the data can also come from the patient, recorded by smart
sensors.
The variety of the data sources have led to an increase in the complexity of the actual data recorded for
each patient.
EHR might have high resolution images from dermatoscopes or other sensors, or low resolution data, such as
manual
clinical observations. This increase in both quantity and quality of the recorded data presents a challenge
for the
clinical practice as well as for medical research to exploit the useful information from such a data set
(Mernik and
Menzies, 2012).
Since HIS consist of a wide range of technologies, such as databases, websites, applications, biomedical,
administrative and financial technologies, technical challenges are arising during the implementation of
such a system.
One of the biggest challenges are the lack of standardization of the digital data and the lack of a
well-developed
healthcare information exchange. During the development and design, the implementers have to spend a
significant amount
of time with ‘cleaning’ and formatting the medical data in order to be able to record it into a central,
common HIS.
Often another technical challenge can arise, which is data sharing over networks. Sharing such data can
lead to
security and data privacy issues. Other obstacles often arise on the management / business level
(Blumenthal, 2009).
Managers, and business owners are typically concerned about the cost and the return of their investment.
This remains a
major issue, as HIS projects typically requires big initial investment, because of the software development
cost. Also,
HIS implementation would typically require huge organizational changes, including buying new devices and
training the
users (Blumenthal, 2009). Finally, as Lapointe and Rivard (2005) pointed out, user resistance to IT is a
novel and
rising issue, and IT resistance researchers need to figure out how and why it occurs, especially in HIS
environment.
Given the high frequency of medication errors, their prevention is a worldwide priority for health systems.
In the UK,
out of 1000 consecutive claims reported to the Medical Protection Society from 1996, 193 were related to
drug
prescriptions. Healthcare information systems, with components such as computerized physician order entry,
bedside
bar-coded medication administration, and electronic medication reconciliation are key to prevent medication
errors
(Agrawal, 2009).
Clinical decision making is a complex process, with several steps that requires the physician to
potentially memorise
and recall a huge amount of data related to a particular patient and drug. Since most of the errors occur
at the
prescription step, a good HIS with a computerized physician order entry (CPOE) is a powerful tool to
improve patients’
safety. A CPOE would make sure that the order is complete and legible, including all necessary information
such as;
prescribed dosage, route of administration and unit doses, patient’s allergies, drug recall warnings and
drug-drug
interactions. It also helps the physician to calculate the correct dosage based on clinical features such
as the weight
of the patient. Another component of a HIS is the bar-coded medication administration (BCMA) system. This
would help to
avoid wrong dosage of medication, or mismatch with the patient’s identity. Previous studies have shown
54-87% reduction
in medication errors where BCMA systems were used (Agrawal, 2009).
HIS can be a great tool to reduces the cost of keeping health records, furthermore it also can improve
workflows,
practice management and pricing. It is expected to permit data sharing between healthcare providers, which
will reduce
office and hospital visits and hospital admissions, and even reduce malpractice lawsuits (Goldschmidt,
2005).
Reducing medical errors and the cost of healthcare are important to increase the overall quality of
healthcare.
However, by harvesting the capabilities of data mining (provided by the huge amount of structured data)
disease
research, prevention, and clinical decision making can be significantly improved (Goldschmidt, 2005).
Opportunities of data mining in healthcare can generally grouped as management of healthcare, customer
relationship
management, detection of fraud and evaluation of treatment effectiveness (Kincade, 1998).
In healthcare management, data mining aims to better identify, track and aid chronic disease states and
patients,
design interventions, and overall reduce the number of hospital admissions. For example data mining can be
used to
group patients by their demographic characteristics and medical conditions to determine which group uses
the most
resources, and develop programs to educate people to prevent or manage their conditions (Kincade, 1998). In
Carnegie
Mellon University, several machine learning approaches were used to detect heart failure in patients. The
university
was working on EHR data, collected by the Geisinger Clinic between 2001 and 2006. By using logistic
regression and
boosting techniques, researchers predicted heart failure more than 6 months before the diagnosis (Wu, Roy
and Stewart,
2010). Data mining also can aid successful infection control or as an automated early-warning system in the
event of
epidemics (Kreuze, 2001). A syndromic system, based on patterns of symptoms, is likely to be more efficient
and
effective than a traditional system that is based on diagnosis. An early warning of the global spread of
the SARS virus
is an example of the usefulness of a syndromic system based on data mining (Brewin, 2003).
The next data mining opportunity option is through the Generally Customer Relationship Management (CRM).
CRM is an
approach at commercial organisations, however the relationship between the patients and the providers is
also
important. As in the case of any commercial organisation, data mining applications can be developed in the
healthcare
industry to determine the preferences, usage patterns, and current and future needs of individuals to
improve their
level of satisfaction (Biafore, 1999). Through the use of data mining, Customer Potential Management
Corporation has
developed a Consumer Healthcare Utilization Index that provides an indication of an individual’s propensity
to use a
specific health care service. The index is generally defined by twenty-five major diagnostic categories,
selected
diagnostic related groups or specific medical service areas. This index (based on millions of healthcare
transactions
of several million patients) can identify patients who can benefit the most from specific healthcare
services. It can
encourage patients who most need a specific care to access it, and continually refine the channels and
messages used to
reach appropriate audiences for improved health and long-term patient relationships and loyalty. The index
has already
been used by OSF Saint Joseph Medical Centre to get the right messages and services to the most appropriate
patients at
strategic times. The result is more effective and efficient communications as well as increased revenue
(Paddison,
2000). Pharmaceutical companies can also benefit from data mining by tracking the physicians and the drugs
prescribed
for patients. It can be checked and evaluated which is the least expensive or most effective treatment plan
for an
ailment, as well as to help identify physicians whose practices are suited to specific clinical trials (for
example,
physicians who treat a large number of a specific group of patients) (Brannigan, 1999).
Data mining applications that attempt to detect fraud and abuse, first establish norms then identify
abnormal patterns
in claims of physicians, laboratories or clinics. A successful example of using data mining to detect fraud
and abuse
was when ReliaStar Financial Corporation has started to use data mining and has reported a 20 percent
increase in
annual savings (Christy, 1997).
Data mining applications can be developed to evaluate the effectiveness of certain medical treatments. By
comparing and
contrasting causes, symptoms, and courses of treatments, data mining can deliver an analysis of which
courses of action
prove effective (Milley, 2000). For example, the outcomes of patient groups treated with different drug
regimens for
the same disease or condition can be compared to determine which treatments work best and are most cost
effective
(Kincade, 1998).
The final category is a more specialized application of data mining in healthcare, such as predictive or
personalised
medicine, analysis of the genome or images produced by neuroscience. There are plenty of examples for
researchers using
machine learning algorithms to find patterns in genetic data to find biomarkers for certain diseases. These
biomarkers
then later can be used to identify disease in potential patients. Boutorh and Guessoum (2016) used a hybrid
machine
learning technique, based on Association Rule Mining, Neural Networks, and Evolutionary Algorithms on
Single Nucleotide
Polymorphism (SNP) data, to classify patients with Autism, Mental retardation, Colon cancer or Breast
cancer. At the
end of the classification, the selected features were identified as possible biomarkers for the particular
disease.
Their Focused Time Delay Neural Network achieved 100% classification result for breast -and colon cancer,
and more than
90% for the mental disorders (Boutorh and Guessoum, 2016).
Another example is Yang et al., 2010, where another hybrid machine learning method was used, on a combined
data of
SNPs, and functional magnetic resonance imaging (fMRI). The selected combined model has achieved 87.2%
accuracy in
predicting schizophrenia in patients (Yang et al., 2010).
As it was earlier discussed earlier, there are several global trends in healthcare, such as moving from
paper based to
paperless systems, or increasing the boundaries of the healthcare system, which allows data sharing across
the
healthcare providers and the patients.
In United Kingdom, National Health Service (NHS) was founded in 1948 with the ideal, that good healthcare
should be
available to everybody regardless of their wealth. NHS always aims to take advantage of the latest
opportunities that
science and technology has to offer, however these changes mean they need to take a longer view called a
Five-Year
Forward View (FYFV) to consider possible opportunities (NHS England, 2017).
Several issues were identified which are in-line with the global trends such as the little amount of time
that
patients, even with long term conditions, actually spend with healthcare professionals. This is mainly
because of the
divide between family doctors and hospitals, and physicians and mental health. This separation makes
patients to visit
multiple facilities with complex conditions, or to bring test results from one physician to another.
NHS as a world leader in primary care computing recognizes the importance of HIS, and the challenges and
opportunities
come with it. However, in the past decades the progress of the development was slower than it should have
been, mainly
because of the two opposite approaches for digital transformation. First, they tried a highly centralised,
national
implementation, however it has failed, because of lack of local engagement, and sensitivity of local
circumstances. The
other approach was to let several independent systems to be developed, however they ended up with several
systems that
don’t share their data.
In the future, the focus will be on systems, that provide a glue between the different parts of the
healthcare systems,
provided they meet the nationally specified interoperability and data standards. To advance the
implementation of the
FYFV, a National Information Board (NIB) has been established, which will publish a set of “road maps” to
support the
transformation. Some of the key elements will include: comprehensive transparency of the data, expanding
the NHS
accredited health apps that patients will be able to use, fully interoperable electronic health records,
family doctor
appointments and electronic prescriptions availability online, and bringing together medical data from
hospitals and
GPs to support quality improvement and research.
Research is vital for NHS, therefore they will continue supporting the National Institute of Health
Research (NIHR) and
the network of specialist clinical research facilities in NHS. Further steps will be taken to speed up
discoveries in
medicine and diagnostics such as reducing the cost of Randomised Controlled Trials (RCTs), expand the Early
Access to
Medicine programme, rolling out new devices and equipment and decommission old ones, and to accelerate the
adoption of
new innovations. To support this acceleration, NHS will develop a small number of “test bed” sites, that
will serve as
a real-world site to test innovations and to help their integration to the existing technology.
Ethical issues mainly raise from three sources, the autonomy of users and privacy. Autonomous patients
expect access to
their data, and many patients will desire some level of control over their records. Such an expectation
conflicts with
the medical and legal utility of a health record and HIS. To tackle this problem, patients should be
restricted to
view, update, but not to delete any information from their health records. The ownership of the records is
another
problem, because given the fact that a health record holds a patient’s private, intimate data, the patients
claim, that
they are the rightful owners. On the other side, companies who develop the HIS, maintain the data storage
and servers,
but even the healthcare providers claim the ownership of the same data. These conflicting interests, must
be rectified
both ethically and legally before a HIS is implemented (Mercuri, 2015).
The privacy of data is important in any information system, however in healthcare it is even more crucial.
Since
patient’s record contains intimate data, once it is revealed, patients can lose trust in the healthcare,
which will
damage the unique doctor-patient relationship. Strict privacy policies, however conflict with data sharing,
therefore a
good public policy has to be defined that protects privacy but in the same time provides a level of
openness that will
benefit the public health (Mercuri, 2015).
For example, a good policy that protects the patient’s privacy and autonomy, would give the patient’s the
right not
just to decide if they wish to share their data, but they could also decide what level of access they want
to guarantee
for which providers, or scientists (Mercuri, 2015).
A large interconnected system, with shared data is a very important part of a good HIS, because of the
great potentials
of the massive amount of raw data, as it was discussed earlier. As part of the system, patients should be
able to
provide access to their data for researchers and scientist, but also, approved scientist should be able to
access data
that was made public (Mercuri, 2015).
In such a system, data security issues arise, especially when portable devices are involved too, therefore
robust
authentication and access control is required, however keeping the balance between making a system secure
or easy to
use is one of the greatest challenge. This is even more difficult, when some of the users might be
physically or
mentally impaired (Cushman et al., 2010).
The necessary non-technical skills will depend on the specific role. For software developers, these skills
are the same
regardless if it’s a healthcare company or a bank. Cognitive abilities such as critical and analytical
thinking or
problem solving is crucial for a programmer. Since most of the companies are working in agile now,
communication skill
became more important. In agile teams, there are regular discussions and “show and tells”, where every
member of the
team (including developers), must stand in front of other people and present their progress, or discuss the
problems.
Some level of business awareness is necessary even for developers, because they should understand the
background of a
project to be able to deliver a good product. With higher positions, such as team leaders and managers,
business
awareness and service delivery management becomes more important. People in these positions are responsible
for a team,
therefore skills like leadership, planning, and time management are very important.
There are several different types of developers in healthcare, and the required technical skills depends on
the type of
the role. In general, employers are looking for multi-talented programmers, with as many skill as possible,
with proven
commercial experience. There are mainly two types of roles: web developers, and application developers. For
web
developers, javascript seems mandatory nowadays, regardless if it is JQuery, NodeJS, Angular, or React.
Html, CSS is
basic for any web developer, C# with .Net, python, or ruby on rails are used at many companies, knowing at
least one of
these technologies can make a big difference.
For application developers, the necessary skills and languages depend on the project that they must work
on. Every
language has the advantages and disadvantages, for example Java is popular because it is platform
independent, and C++
is usually used when the focus is on the performance. Python became more popular recently, mainly because
it is easy to
learn, and it has many libraries that are used in machine learning. One of these software libraries is
Tensorflow
developed by Google’s Brain Team. Tensorflow is a very powerful, open source library, that can be used in
research and
production without rewriting any code. The flexible architecture allows the developers to deploy the
computation to one
or more CPUs or GPUs in any environment (server, desktop, mobile) with a single API. Knowing python and /
or Tensorflow
is a great advantage for any developer who wish to work in data science.
Stephen Docherty is a Chief Information Officer (CIO) for South London and Maudsley (SLaM) NHS Foundation
Trust. SLaM
provides mental health services and substance misuse services in the community of Lambeth, Southwark,
Lewisham, and
Croydon. It is the only NHS Trust with a biomedical research centre, hosted in King’s College. Besides the
health
services, SLaM also provides a range of education, training and learning opportunities (Slam.nhs.uk,
2017).
Stephen, during his expert talk introduced us an innovative IT strategy, that was signed off at SLaM in
March 2015.
Some of the key points of the new strategy were: “cloud first” approach, service management, frameworks,
and culture
change.
Cloud based services played a key point in the transformation. During the past two years, SLaM migrated all
staff to
Microsoft Office 365, which is the cloud based version of Microsoft Office. This helps the support of the
users, since
the application is cloud based, therefore the maintenance is Microsoft’s responsibility. As a business
intelligence
solution, Microsoft PowerBI has been introduced, which is another Microsoft cloud based application
providing modern
data visualization and analytical tools. More than 2000 old devices were also replaced, and the electronic
health
record system called electronic Patient Journey System (ePJS) has been tested on mobile devices. In 2013,
SLaM released
myHealthLocker with the purpose of engage patients online in their care, however the take up was poor. Now,
a new
version of the application (HealthLocker V2) is being developed and tested through a user-led design
process, focused
on delivering a clinically useful tool. Through the application, patients will be able to send
confidential, secure
messages to the staff, access their care plan, or even track their sleep. The development team is using an
agile
development process, with involving service users, carers, clinicians, and researchers to make sure that
the
application will meet their needs. HealthLocker is going to be launched by the end of July, but the
development won’t
stop, it is planned to add new features in the coming months (Healthlocker.uk, 2017).
As discussed earlier, the successful adoption of digital and technical innovations, requires active support
of service
users. SLaM created a forum called SL@M Connect, which is a weekly meeting between clinicians, digital
services and
communications team to discuss the issues, inform people about digital developments, and to discuss new
ideas.
This analysis has looked at future of information technology, and how it relates to healthcare. British
trends in
healthcare are following the global trends, which was proven by looking at the Five Forward Year View of
NHS. The
opportunities with healthcare information systems are certainly outweigh the challenges, however to make a
successful
digital transformation that satisfies all stakeholders and users of the system, the designers and project
leaders must
achieve a cultural change, and they must involve every user in the design process. Such systems also raise
several
security, data privacy, and piracy issues, that has to be taken very seriously, because of the nature of
the data.
Despite of these challenges, digital innovation is already happening in healthcare, one great example is
CIO Stephen
Docherty at South London and Maudsley NHS Foundation Trust. Stephen and his team, restructured the IT
department and
rolled out several products that will help data sharing, business intelligence, and remote patient care,
and the
foundation also has a biomedical research centre where data mining algorithms and tools are used on the
collected data
for research purposes.
Agrawal, A. (2009). Medication errors: prevention using information technology systems. British Journal of
Clinical
Pharmacology, 67(6).
Bellman, R. E. (1978). An introduction to artificial intelligence: can computers think? San Francisco: Boyd
& Fraser
Pub. Co.
Biafore, S. (1999). Predictive solutions bring more power to decision makers. Health Management Technology,
20(10).
Blumenthal, D. (2009). Stimulating the adoption of health information technology. The New England Journal
of Medicine,
360(15).
Brannigan, M. (1999). Quintiles seeks mother lode in health “data mining.” Wall Street Journal, 2 (1).
Brewin, B. (2003). New health data net may help in fight against SARS. Computerworld, 37(17).
Boutorh, A. and Guessoum, A. (2016). Complex diseases SNP selection and classification by hybrid
Association Rule
Mining and Artificial Neural Network—based Evolutionary Algorithms. Engineering Applications of Artificial
Intelligence, 51.
Christy, T. (1997). Analytical tools help health firms fight fraud. Insurance & Technology, 22(3).
Cushman, R., Froomkin, A., Cava, A., Abril, P. and Goodman, K. (2010). Ethical, legal and social issues for
personal
health records and applications. Journal of Biomedical Informatics, 43(5).
Emc.com. (2017). Executive Summary: Data Growth, Business Opportunities, and the IT Imperatives | The
Digital Universe
of Opportunities: Rich Data and the Increasing Value of the Internet of Things. [online] Available at:
https://www.emc.com/leadership/digital-universe/2014iview/executive-summary.htm [Accessed 9 Jul.
2017].
Goldschmidt, P. G. (2005). HIT and MIS: implications of health information technology and medical
information systems.
Communications of the ACM, 48(10).
Haux, R., Winter, A., Ammenwerth, E. and Brigl, B. (2004). Strategic Information Management in Hospitals.
New York, NY:
Springer New York.
Healthlocker.uk. (2017). Healthlocker. [online] Available at: https://www.healthlocker.uk/ [Accessed 22
Jul. 2017].
Kincade, K. (1998). Data Mining: Digging for Healthcare Gold. Insurance & Technology, 23.
Khosrowpour, M. (2015). Encyclopedia of information science and technology. 3rd ed. Hershey: Information
Science
Reference.
Kreuze, D. (2001). Debugging hospitals. Technology Review, 104(2).
Kurzweil, R. (1990). The age of intelligent machines. London: MIT Press.
Mercuri, J. (2015). The Ethics of Electronic Health Records | Clinical Correlations. [online]
Clinicalcorrelations.org.
Available at: http://www.clinicalcorrelations.org/?p=2211 [Accessed 18 Jul. 2017].
Mernik, M. and Menzies, T. (2012). 2012 First International Workshop on Realizing AI Synergies in Software
Engineering.
1st ed. NJ, USA: Press Piscataway.
Milley, A. (2000). Healthcare and data mining. Health Management Technology, 21(8).
Moravec, H. (1998). When will computer hardware match the human brain? Journal of Evolution and Technology,
Vol.
1(1541-0099).
NHS England, N. (2017). NHS England » NHS Five Year Forward View. [online] England.nhs.uk. Available at:
https://www.england.nhs.uk/five-year-forward-view/ [Accessed 12 Jul. 2017].
Paddison, N. (2000). Index predicts individual service use. Health Management Technology, 21(2),
14-17.
Rich, E. (1991). Artificial intelligence. New York: McGraw.
Slam.nhs.uk. (2017). Home - South London and Maudsley NHS Foundation Trust.. [online] Available at:
http://www.slam.nhs.uk/ [Accessed 17 Jul. 2017].
Turing, A. (1950). Computing Machinery And Intelligence. Mind, LIX(236).
University of Reading. (2017). University of Reading. [online] Available at:
http://www.reading.ac.uk/news-and-events/releases/PR583836.aspx [Accessed 9 Jul. 2017].
Wu, J., Roy, J. and Stewart, W. (2010). Prediction Modeling Using EHR Data. Medical Care, 48.
Yang, H., Liu, J., Sui, J., Pearlson, G. and Calhoun, V. (2010). A Hybrid Machine Learning Method for
Fusing fMRI and
Genetic Data: Combining both Improves Classification of Schizophrenia. Frontiers in Human Neuroscience,
4.
Zwolenski, M. and Weatherill, L. (2014). The Digital Universe Rich Data and the Increasing Value of the
Internet of
Things. Australian Journal of Telecommunications and the Digital