Challenging Assumptions on Big Data

Challenging Assumptions on Big Data

This is part of a Penton Technology special series on big data.

As you read through Penton Technology’s special report on big data and the opportunity it offers you as a service provider, you will learn how big data can be transformative if used properly. It’s not just marketers that can take advantage of all the insights big data can provide; governments, healthcare systems and others in the public arena can use big data to better serve their constituents by providing adequate healthcare and housing, for example.

As technologists, we often talk about big data as a technical challenge: how will we be able to support the growth of data? Will our clouds be able to scale with the data deluge? How will we secure all of this data? It can be easy to forget about the other side, which deals with the ethics of big data.

Discrimination and Big Data

The White House has actually been exploring this side of big data through the Obama Administration’s Big Data Working Group in hopes of creating dialogue around some of these issues. The group has released a series of reports on big data; the most recent report, released last month, looks at the assumption that big data techniques are unbiased.

“The algorithmic systems that turn data into information are not infallible—they rely on the imperfect inputs, logic, probability, and people who design them. Predictors of success can become barriers to entry; careful marketing can be rooted in stereotype. Without deliberate care, these innovations can easily hardwire discrimination, reinforce bias, and mask opportunity,” according to a blog post.

Some examples of these biases could leave underserved and low-income families out of credit and employment opportunities. The Federal Trade Commission said that to maximize the benefits and limit the harms of big data, companies should consider how representative their data set is, if their data model accounts for biases, the accuracy of their predictions based on big data, and whether their reliance on big data raises concerns around fairness.

“Ideally, data systems will contribute to removing inappropriate human bias where it has previously existed. We must pay ongoing and careful attention to ensure that the use of big data does not contribute to systematically disadvantaging certain groups,” the report said. “To avoid exacerbating biases by encoding them into technological systems, we need to develop a principle of ‘equal opportunity by design’—designing data systems that promote fairness and safeguard against discrimination from the first step of the engineering process and continuing throughout their lifespan.”

The White House report goes through several case studies and recommends that “public and private sectors continue to have collaborative conversations about how to achieve the most out of big data technologies while deliberately applying these tools to avoid—and when appropriate, address—discrimination.” The report offers 5 recommendations:

  1. Support research into mitigating algorithmic discrimination, building systems that support fairness and accountability, and developing strong data ethics frameworks.
  2. Encourage market participants to design the best algorithmic systems, including transparency and accountability mechanisms such as the ability for subjects to correct inaccurate data and appeal algorithmic-based decisions.
  3. Promote academic research and industry development of algorithmic auditing and external testing of big data systems to ensure that people are being treated fairly.
  4. Broaden participation in computer science and data science, including opportunities to improve basic fluencies and capabilities of all Americans.
  5. Consider the roles of the government and private sector in setting the rules of the road for how data is used.

Big Data Algorithms Can Learn Discrimination

Algorithms learn from the external process of bias or discriminatory behavior, according to a blog post by The Ford Foundation.

“The origin of the prejudice is not necessarily embedded in the algorithm itself: Rather, it is in the models used to process massive amounts of available data and the adaptive nature of the algorithm. As an adaptive algorithm is used, it can learn societal biases it observes.”

In order to address this, policymakers must learn more about big data, and algorithms that underpin “critical systems like public education and criminal justice” must be transparent. There also needs to be updated regulation around the use of personal data, according to the report.

Data as a Liability

A report by our sister site Windows IT Pro looks at the amount of responsibility companies take on when they store big data, referencing a report by Quartz which states data companies will not own data, “they will just manage the flows of data between those that have data and those that need it.”

The conversation around discrimination and big data challenges the idea that big data will replace human decision-making. Big data algorithms can assess, sort, and analyze all sorts of data, but without a human ensuring that the data is being used properly, there could be major consequences.

Source: TheWHIR

Drowning in Data: How Channel Providers Help Customers Make Sense of the World of Information Around Them

Drowning in Data: How Channel Providers Help Customers Make Sense of the World of Information Around Them

Whether or not you realize it, you are contributing to the 2.5 exabytes of data that is produced around the world every single day.

You start creating data when you wake up if you’re one of the 80 percent of smartphone users who check their phones before they brush their teeth. You may even be creating data overnight as you wear your smart watch that collects information on your sleep habits.

So what is being done with that data? It’s a good question, and just one piece of the massive jigsaw puzzle that is making sense of big data.

In order to break it down for you, Penton Technology’s Channel brands The VAR Guy, MSPmentor, Talkin’ Cloud and The WHIR have focused this month on big data and the data deluge.

On Talkin’ Cloud, we share the 7 big data jobs that you must know about. Demand for these jobs is going to continue to grow, so gaining skills in big data is a great way to future-proof your tech career.

On The VAR Guy, Editor in Chief Kris Blackmon shares her thoughts on the coolest data tools of 2016 so far. Definitely worth scoping out if you’re in the market for some new ways to wrangle all of that data.

We’re breaking it down even further with a gallery on 7 open source big data analytics and storage tools by Christopher Tozzi, our resident open source expert.

Also on The VAR Guy, we test your knowledge of big data with an interactive quiz. (We’re sure that by the time you’ve read our entire editorial package on big data, you’ll be on your way to expert status.)

On the WHIR, Clarity Channel Advisors Chief Advisor and HostingCon Evangelist Jim Lippie describes how business analytics, business intelligence and big data intersect to bring you real business insights – regardless of the size of your business.

By 2020, the total amount of data around the world will reach 44 zettabytes. That’s a lot of data to go around. Our hope is that with this editorial package, you will walk away with the knowledge to use big data to bring success to your organization.

Source: TheWHIR

HTBase Named A Cool Vendor In Hyperconvergence By Gartner

HTBase Named A Cool Vendor In Hyperconvergence By Gartner

HTBase has announced it has been included in the list of “Cool Vendors” in the April 27, 2016 report by Gartner, Inc., Cool Vendors in Hyperconvergence, 2016. 

Enterprises are modernizing their data center and cloud practices with HTBase. 

“We are the only platform that allows organizations to easily build and deploy a modern Cloud Enterprise Infrastructure — consisting of containers, multiple virtualization technologies, storage, network, real-time analytics, sef-service and more — on a single platform. We are proud to have been selected as a Gartner Cool Vendor,” said Louis Smith, partner manager, HTBase.

HTBase differentiates itself in the hyperconvergence space by delivering on the concept of a true software-defined data center (SDDC) infrastructure, while enabling IT to retain existing infrastructure or develop new hardware platforms in greenfields. I&O leaders and decision makers can, therefore, retain and adapt their existing x86 hardware and storage for hybrid cloud solutions, while delivering IT as a service (ITaaS) to their business units. 

HyperTask, the base of our Hyperconvergence platform, centrally manages multiple hypervisors, including open-source kernel-based virtual machine (KVM), ESXi, HyperV and Xen; enables existing VMs to migrate between different hypervisors; and supports Docker, bare metal, and integration to AWS and Azure. It also manages storage, network and compute resources, with capacity management, storage resilience, high availability, striping, deduplication, snapshot/replication (with resulting disaster recovery) and automation tools through REST APIs. 

Since launching, HTBase has helped CIOs and I&O leaders designing infrastructure as a service with a strong cloud strategy. Moreover, our platform has helped I&O leaders to migrate from legacy environments (regardless of hypervisor or bare metal) to a centrally managed — but independently run — business units, controlling their own resource allocation by either conventional VMs or by containers.

Source: CloudStrategyMag

Ecommerce as an Opportunity for Service Providers

Ecommerce as an Opportunity for Service Providers

The program of educational sessions at HostingCon Global 2016 New Orleans has been carefully constructed to cover the full range of challenges and opportunities for business success in the web hosting and cloud provider ecosystem. Ecommerce is one of the fastest growing and potentially most lucrative opportunities currently in the industry, but many companies have been slow to adapt to take advantage of it.

US online retail sales are estimated at $340 billion in 2015, and a report by Google and PayPal estimated the four largest online retail markets will reach $500 billion within two years. Google and Forrester have predicted the ecommerce market in India will reach $15 billion this year as well, and new ecommerce markets are opening at the speed of internet access improvement.

[Get Your HostingCon Global Exhibit Booth Now Before it’s Too Late]

That’s why one of the sessions in the Marketing Track of HostingCon Global is “How Service Providers Can Take Advantage of the Booming E-Commerce Market.” Wilfried Beeck, CEO of ePages, will lead an exploration of the role of web hosts, ISPs, and service providers in ecommerce, particularly for SMBs. Competition among ecommerce companies grows constantly, but few of them deal directly, let alone effectively with SMBs.

Beeck founded ePages in 1983, after studying Mathematics and Computer Sciences. He also co-founded Intershop Communications GmbH, and managed its IPO in 1998. He is currently involved with a number of other high-tech companies as an investor and board member.

His expertise will be highly valuable to many hosts, and just may be the start of a big break for your web hosting company at HostingCon Global 2016. Register by June 10 to save $100 with the Early Bird rate.

Source: TheWHIR

The Power and Profitability of Women in Tech

The Power and Profitability of Women in Tech

A 2016 Peterson Institute study concluded that inclusion of women in leadership roles has a positive impact on company profits. It surveyed 21,980 companies globally across 91 countries and found that companies that had at least 2 women in executive positions in the C-Suite and 2 women in board positions actually had higher profitability than those that didn’t have this diversity. The highest statistical correlation appeared in the C-Suite executive roles.

Research shows that a gender-diverse workforce leads to a more inclusive organizational culture which has incredible power to build increased profits for organizations. Diversity and inclusion have become a powerful tool in a company’s business strategy for success.

HostingCon education has always focused on new ideas to increase profits and revenue for the cloud and service provider ecosystem. To that end, the Internet Infrastructure Coalition is sponsoring the panel, “Connecting With Regional Groups To Improve Your Company’s Diversity” moderated by hosting industry leader Elizabeth Kurek, Director of Channel Marketing, Virtuozzo.

A study from Catalyst.org, a known authority on women in business, cites four qualities of being an inclusive leader: empowerment, accountability, courage, and humility. Being an inclusive leader means you create a workplace culture where employees feel connected and supported. Inclusive leaders foster innovation and creativity, and an opportunity for everyone to advance and thrive within the organization. Are you an Inclusive Leader? Here are the attributes:

Are you an Inclusive Leader? Here are the attributes:

Empowerment – Inclusive leaders encourage team members to solve problems, not just do a job. They want their team to have new ideas and develop skills that will empower them to be a better resource for the whole organization.

Accountability – Inclusive leaders are confident in their team. They don’t micro-manage but, instead, hold them responsible for their performance. They let them own the responsibilities and have control over the outcomes.

Humility – Inclusive leaders are open to input and feedback. They can admit mistakes and learn from them. They will know their own limitations and seek assistance from others to overcome and be successful.

Courage – Inclusive leaders will put aside their personal interests to achieve what needs to be done. They are not afraid to do what is right, and will often take some risks to ensure the correct outcome.

An inclusive leader will foster a workplace where diversity is valued, and where women can find a balanced organization. However, a single inclusive leader can’t change an entire organization. There needs to be top-down sponsorship at the senior level in organizations to really drive and sustain a culture of inclusion and a balanced workforce.

Organizations need to build a culture in teams that celebrates the collective brilliance of the organization. Successes and failures should be celebrated. Cultures that embrace innovation, experiment with pilots and new ideas and improve along the way will create an atmosphere for success.

There are several things specific things any organization can do to build a culture of inclusion.

Attract Female Candidates – One of the simplest things you can do is targeted recruitment programs to have more women candidates for open roles in the organization. Being deliberate in the organization about hiring positions that can have inclusion benefits is important.

Identify Internal Talent – Be on the lookout for women to mentor into roles with additional responsibility and cross functional training. Identifying talent early and developing them inside the organization will greatly benefit your efforts of inclusion.

Enable Women into the Executive and Board level – Building women into management roles, hiring at least 2 executives into your C-suite, and ensuring your corporate or advisory board has at least 2 women are all techniques to enable an inclusive and ultimately more successful organization.

Here are a few resources to enable the success of tech men and women together in the workplace:

Global cause for social good at http://heforshe.org/en

The authoritative agency on women: http://www.catalyst.org

The 20 by 2020 initiative – 20% women in America on public company corporate boards https://www.2020wob.com

Training for women executives, women in business, mentoring forums: www.ccle.org

Growing program to teach middle school girls to like and value technology (when they learn and love it at this age, they are more likely to enter the field later) : www.techgirlz.org

Inclusion is a trait to be valued throughout an entire organization. Being an inclusive leader and taking proactive steps to be inclusive as an organization are great ways to help promote women in tech. Done well, any organization can find its way to increased success and higher profitability by employing these techniques.

Source: TheWHIR

Micro-Segmentation to Your Cloud

Micro-Segmentation to Your Cloud

Security for public and community clouds has greatly improved recently. Major cloud service providers such as Amazon Web Services and Microsoft have made substantial investments in security to help ensure their subscribers’ data is safe and their cloud experience is exceptional. In fact, the security from such cloud service providers is better than in many companies’ own data centers. Some companies now actually view security as one of the primary justifications, in addition to cost savings and agility, for moving their applications to the cloud.

Still, the thought of putting critical data on the cloud accessible by just about anyone is really scary and the perceived added vulnerability is preventing many enterprises from fully leveraging the cloud. They may be using the cloud for non-critical data such as test and development but not for their core business applications that access their most sensitive data.

New micro-segmentation offerings are now available to provide enterprises with added layers of cloud security and instill the confidence they need to put more of their enterprise applications on the cloud. In doing so, large companies have the opportunity for tremendous cost savings, to make their products and services more globally accessible and to dynamically adjust to business conditions in real time.

There are five security advantages these new micro-segmentation offerings provide to the cloud that haven’t previously been obtainable:

  1. New micro-segmentation offerings enable companies to use a consistent set of tools for both their local data centers and the cloud. In the past, security administrators and operators have had to use different security tools for their local data centers and each cloud service provider. This meant the security roles and policies within their companies had to be mapped and maintained across multiple data centers (private and public) and different toolsets – a very complex and costly undertaking. Contemporary micro-segmentation tools work within a company’s local data center and across leading cloud service providers’ data centers, greatly simplifying operations and reducing costs.
  2. New micro-segmentation technologies provide encryption within the cloud from virtual machine to virtual machine. Traditionally, it was believed that data was secure enough once within the security protection provided by the perimeter of the public data center. New micro-segmentation technologies now encrypt data between the virtual computing instances within the cloud for each company and user community. This encrypted added layer of security is a major step forward for protecting the most sensitive data in the cloud.
  3. New micro-segmentation technologies use concealment as a basis for security strategy. Traditional security offerings have used a fortress strategy in which they build a wall around the data they are trying to protect, but still leave the processing environment visible to cybercriminals. New micro-segmentation offerings conceal the applications and data on the cloud to the point where it doesn’t appear like anything is active. Cybercriminals aren’t seeing information such as active IP addresses, operating system types or other information that would provide insight as to how to attack a company’s data. The cybercriminals are not aware of the applications and data in the cloud, and therefore it is not a target.
  4. Micro-segmentation prevents lateral movement of security infiltrations to the data center. A key concern of the cloud is that a company’s data may be compromised because “the other guy” on the cloud didn’t take the appropriate precautions. Micro-segmentation protects cloud users from “the other guy.” Each company’s cloud processing environment is isolated and secure. Malware and cybercriminals are bounded by any given segment, and this makes all users of the cloud more secure.
  5. Micro-segmentation can prevent security breaches in the cloud. A security breach occurs only when a company’s data leaves the cloud, not when the actual security infiltration occurs. New micro-segmentation capabilities contain threat intelligence that recognizes abnormalities and invoke policies to prevent data from leaving the cloud where these inconsistencies occur until the appropriate security analysis can be applied.

The cloud has never been more secure and micro-segmentation takes cloud security to amazing to new levels, so that enterprises can fully realize the benefits of the cloud even for their most critical and sensitive data.

About the Author

Rod Head shotRod Sapp, Global Vice President, Security and Cloud Product Management, Unisys

Rod Sapp manages the Unisys products portfolios including Enterprise Servers, storage, security products and cloud products.

Rod and his team are responsible for business planning, server portfolio definition, partner selection & management, and product launch and lifecycle management.

Rod’s organization collaborates extensively with Unisys sales channels and clients as well as Unisys engineering and services organizations to develop market-leading whole product solutions. Rod’s organization works closely with Unisys strategic technology partners including Dell, IBM, EMC, Oracle, VCE, Intel and Microsoft.

Rod’s background includes ten years of sales experience and twenty years of product and portfolio management that has included the Unisys ClearPath mainframe servers and more recently comprises the Unisys X86 business, storage offerings and software products.

Source: TheWHIR

Meeker: Internet Trends Changing Faster Than Ever

Meeker: Internet Trends Changing Faster Than Ever

India has passed the US as the second largest global internet market, behind China, with 277 million users with an annual growth rate of 40 percent, according to the latest research released by Mary Meeker of Kleiner Perkins Caufield & Byers (KPCB). The Internet Trends 2016 presented at Vox Media’s Code Conference on Wednesday shows major change happening in advertising methods, web search, and internet video viewing.

The report comes as DigitalOcean has opened up its second Asian data center in India, one of many US-based internet companies that has expanded to the country in recent months.

SEE ALSO: IoT to Drive Next Wave of Connected Devices: Report

It is little surprise that internet advertising is growing, or that mobile advertising growth outpaces desktop, but the extent of these trends is striking. There was a 20 percent increase in overall spending to $60 billion in the US in 2015, and mobile ad spend roughly doubled in that time. Still, the report shows that based on time spent, mobile still represents a major opportunity for advertisers, with 25 percent of time but only 12 percent of advertising spent on mobile. Non-mobile internet use takes 22 percent of consumer’s time, and 23 percent of advertising spending.

Ad blockers and muted videos represent major growing barriers to the effectiveness of that advertising, however, with 93 percent of internet users considering using ad blocking software, and 81 percent muting video ads. Privacy is also a looming issue, with 50 percent saying they are very concerned about data privacy and how companies use customer data.

READ MORE: China Said to Push for More State Control Over Video Websites

US data privacy regulation policy is in currently in development by the FCC, with advertisers expressing fear earlier this week that their effectiveness could be curtailed by proposed rules.

The report also highlights the declining use of text as a form of internet communication, suggesting it is being replaced in many areas by voice and image. The number of Baidu queries by speech, for instance, had reached one in ten in September 2014, when the company’s chief scientist Andrew Ng predicted that half of all searches will be done by voice or image in 2020.

As big data is adopted by businesses, opportunities for business apps will continue to emerge, according to the report, which is potentially a boon for both web hosts and cloud service providers.

A general point made throughout the report is that the pace of market transformation related to the internet is out of proportion with historical changes in media or technology. The rapid pace has made many internet companies highly profitable very quickly, but it puts pressure on companies to be nimble and forward-thinking.

Source: TheWHIR

Dataguise Survey Highlights IT Security Issues

Dataguise Survey Highlights IT Security Issues

Dataguise has announced the findings of a new survey titled “Strategies for Securing Sensitive Data.” In the survey, 100 senior IT decision makers, including CxOs, VPs, directors, and managers were questioned on the topic of sensitive data security, including technologies in use, impacts to businesses when failures occur, and accountability after such events. The survey participants represented firms from a wide variety of industries that were chosen for the intensity at which they consume data. Conducted between March and April of 2016, the survey uncovers several truths about sensitive data management, risks, and increasing budgets for improving IT security infrastructure.

In March of 2016, Dataguise commissioned Gatepoint Research to conduct an invitation-only survey of enterprise IT executives regarding strategies for securing sensitive data. Candidates were chosen from a wide range of industries, including financial services, health care, manufacturing, business services, consumer services, retail, media, and education. 54% of those that completed the survey work for Fortune 1000 organizations with revenues over $1.5 billion. 20% work for medium to large firms whose revenue is between $250 million and $1.5 billion. 26% are employed by small enterprises with less than $250 million in revenue.

Observations and conclusions of the 13 question survey included the following:

  • Companies are transitioning toward big data frameworks, including cloud-based environments such as Microsoft Azure HDInsight. 28% of respondents report more than a year of experience with these big data repositories and another 38% in various stages of adoption.
  • Data security challenges often have a negative impact on organizations with 73% reporting that data security concerns terminate or delay data-driven business initiatives.
  • Companies use multiple security solutions to protect sensitive data, with 82% using network monitoring, 80% leveraging data encryption, 79% implementing access controls, 69% installing perimeter controls, 63% using volume and file encryption, and 43% implementing data masking.

Even with multiple layers of security in place, less than half of all respondents did not believe that their data was secure with only 47% of respondents confident that their sensitive data throughout their organization was safe. Furthermore, it was revealed that sensitive data within organizations can be widely accessed by a large number of individuals. In addition to 80% of respondents indicating that their IT teams had access to sensitive data, 40% said test and development teams also had access and 29% indicated that end-users throughout the enterprise maintained the ability to view this information. Finally, while 62% of those surveyed said their firms passed security audits, 11% failed and 20% were unclear if they passed their audit or not.

Identifying where the buck stops when unauthorized access to sensitive data occurs, the survey also asked who would be held accountable if the organization encountered a breach. 88% of respondents said that their IT security team (including the CISO/CIO) would face scrutiny. 47% said their CEO or board of directors would be placed with the responsibility. 38% of organizations would point to the chief data officer (CDO) for the breach and 24% would fault the user or users who created the data. The takeaway here is that IT security teams are at the greatest risk should a situation occur and must strengthen their data infrastructure to ensure the danger of unauthorized access remains low.

“As we have experienced, many companies are throwing everything they have at IT security challenges. The problem is that even multiple point solutions still leave gaps that put these organizations at risk,” said JT Sison, vice president of marketing and business development for Dataguise. “Addressing this at the data layer plugs the remaining gaps, regardless of its migration across systems and networks. Additionally, platform agnostic monitoring of this sensitive data provides precise intelligence to administrators, providing a much higher level of protection for greater levels of confidence.”

A complete copy of survey results are available for free download at the website.

Source: CloudStrategyMag

IoT to Drive Next Wave of Connected Devices: Report

IoT to Drive Next Wave of Connected Devices: Report

The Internet of Things (IoT) will surpass cell phones as the largest category of connected devices in 2018, according to research released this week by Ericsson.

The Ericsson Mobility Report covers a variety of topics related to mobile connections, including the current explosion in video streamed to mobile devices.

Ericsson estimates there will be roughly three billion IoT devices in North America alone by 2021. In the same year, there will be 450 million mobile subscriptions and mobile data traffic will reach 22 GB per smartphone each month.

SEE ALSO: Special Report: The Internet of Things and You

Mobile traffic growth will be driven by streaming video, as consumers now in their teens lead a shift away from traditional televisions. Smartphone TV and video viewing by teens has increased by 85 percent in the last 4 years. The shift will see streaming video, including embedded video in social media and on web pages, grow by 55 percent to well over 30 exabytes globally each month, or close to 70 percent of all mobile data traffic, over the next five years. Web browsing by contrast will grow at a compound annual rate of 15 percent.

LTE subscriptions increased by 150 million to 1.2 billion in Q1 2016. Smarphone subscriptions also rose, and Ericsson expects them to exceed basic phone subscriptions in the third quarter of this year.

The report also calls for “global spectrum harmonization to secure early 5G deployments.”

Cisco also predicts that video will make up more than half of all mobile data traffic by 2020, and Gartner identified mobile video as a huge opportunity for communication service providers (CSPs) last year. Data center interconnection is another market likely to benefit from mobile video growth.

Ericsson partnered with AWS in February to encourage telecoms and CSPs to build links to the cloud to improve the efficiency of mobile app delivery, in anticipation of increased data demands. The company also joined with Orange Telecom in November to test different networking approaches and hardware for IoT devices.

Source: TheWHIR

Had it with Apache Storm? Heron swoops to the rescue

Had it with Apache Storm? Heron swoops to the rescue

Last year, Twitter dropped two bombshells. First, it would no longer use Apache Storm in production. Second, it had replaced it with a homegrown data processing system, Heron.

Despite releasing a paper detailing the architecture of Heron, Twitter’s alternative to Storm remained hidden in Twitter’s data centers. That all changed last week when Twitter released Heron under an open source license. So what is Heron, and where does it fit in the world of data processing at scale?

A directed acyclic graph (DAG) data processing engine, Heron is another entry in a very crowded field right now. But Heron is not a “look, me too!” solution or an attempt to turn DAG engines into big data’s equivalent of FizzBuzz.

Heron grew out of real concerns Twitter was having with its large deployment of Storm topologies. These included difficulties with profiling and reasoning about Storm workers when scaled at the data level and at a topology level, the static nature of resource allocation in comparison to a system that runs on Mesos or YARN, lack of back-pressure support, and more.