Latest Transparency Software News
Jul 3, 2013
- David Floyer Wikibon recommends that CIOs put in place a four-point plan for implementing an all-flash datacenter strategy: 1/ Migrate to an All-flash Datacenter by 2016 2/ Proof of concept for Developer Data Sharing - Expect 2 x Productivity & 3 x Code Quality 3/ Reorganize Data Center Operations around Principle of Application Data Sharing 4/ Use IT Budget Savings to Fund Systems of Intelligence This research provides a technology understanding, the business case, and the strategic first steps on the road to an all-flash datacenter. - David Floyer IT will sound like yesterday’s news if it continues to advocate tiering, advocate putting the most important data on high performance storage, and declaring that less expensive lower performing storage is key and necessary. Today’s flash reality is that a single compressed/de-duplicated physical version of data must be shared between as many different applications as possible, because the incremental cost of creating and sharing another copy is close to zero. - George Gilbert Systems of Intelligence will power the next generation of enterprise applications built on big data. Line of business executives charged with digital transformation must understand the dynamics of Systems of Intelligence well-enough to be effective sponsors of new systems. IT executives must be effective partners and understand how to build Systems of Intelligence on top of Systems of Record. They will also need to understand how to build a radically new infrastructure to support these systems. - Stuart Miniman OpenStack is now positioned as an Integration Engine - not the entire stack, but like Linux, could grow into a critical component of IT over the next decade. This article examines the real state of OpenStack after the Kilo release and 2015 OpenStack Summit in Vancouver. - David Floyer Wikibon believes CIOs should put the highest priority in developing an all-flash datacenter strategy, and implement it so that all storage acquired and deployed is 100% all-flash by 2016. The key short-term focus for CIOs is to ensure that data sharing is optimized, and that IT organization and objectives support and drive increases in data sharing. The key long-term focus is to release the potential of the best and brightest in the organization to imagine an organization supported by applications without data limits. - Stuart Miniman The emerging PaaS platforms (Cloud Foundry, Heroku, OpenShift, Docker) are all evolving to capture the mindshare of developers. While the PaaS battles are still in early days, Cloud Foundry is emerging as a leading open source platform and ecosystem to build and deploy the new Cloud Native applications that are driving these economic changes via software. But this Platform-as-a-Service (PaaS) model requires changes to technology, skills (people) and internal IT processes. This means enterprises need to become familiar with this development and operational model to best understand how they can accelerate using Cloud Native applications to give themselves a business differentiation. - David Floyer Platform-as-a-Service (PaaS) is a full application lifecycle cloud service including initial development, testing, deployment, operations and maintenance. Wikibon defines three different PaaS cloud services, 1/ PaaS integrated with IaaS, an integrated platform (e.g., IBM Bluemix, EMC Pivotal Cloud Foundry), 2/ PaaS on top of IaaS, combining development services onto a specific platform (e.g., AWS, Microsoft Azure) and 3/ PaaS on top of SaaS, a development front-end to a SaaS platform with its underlying infrastructure (e.g., ServiceNow, SalesForce). Wikibon will be using this topology to define its PaaS analysis and forecasts. - Jeff Kelly When I began covering the Big Data market for Wikibon back in early 2011, it was early days. Big Data technologies such as Hadoop and NoSQL, while still early in their development, were becoming better known by enterprise practitioners. The market for commercial Big Data technologies and services was small but on the verge of rapid growth. But from a business use case perspective, Big Data was long on promise but short on specifics. A lot has changed over the last four years. Without question, Big Data technologies have developed at breakneck speed due in large part to a vibrant open source community of developers. Hadoop in particular has taken great strides. What was once a batch processing, unsecure and somewhat finicky framework is now a much more comprehensive, enterprise-grade, multi-application supporting Big Data platform. - Steve Chambers Hybrid cloud is being touted as the best solution for enterprises today but it is under attack. Complex enterprises have complex needs which means they are unlikely to get everything they need from one hybrid cloud. Investing in a single hybrid cloud to glue one private and one public cloud together is not the ultimate answer, as recognised by many referring to hybrid cloud as "a journey". However, the danger is that if you design for a dual-element hybrid cloud then you will not equip yourself for the extensibility required to snap in/out new/old cloud elements over time, and you will still be down a dead end. - Jeff Kelly MarkLogic held its annual customer and partner conference in San Francisco earlier this month. The company, whose NoSQL database evolved from a simple XML data store and is beginning to disrupt the traditional RDBMS market, is experiencing solid growth. For the calendar year 2014, the company reported total revenue in excess of $100 million for the first time in its 14-year history. It is the largest of the NoSQL pure-play vendors and recently added JSON support, semantic and bitemporal analytics capabilities, and incremental back-up capabilities to its flagship database.