Data Modeling Articles
Open source databases like MySQL, MongoDB, Cassandra, PostgreSQL and others have gained wide adoption in recent years. However, not all companies know if open source is right for them or how to migrate their workloads over. DBTA recently held a webinar with Joe McKendrick, research analyst, Unisphere, and Nihal Mirashi, principal product marketing manager, Pure Storage, who discussed why storage matters and the consequences of mismatched storage options, along with how modern storage can maximize application performance and scalability.
Posted November 18, 2020
Revelation is releasing a "Roll Up" patch to OpenInsight version 9.4 that will enhance debug functionalities and improve operations. This patch was made as a result of user reports and suggestions, said Mike Ruane, president and CEO of Revelation.
Posted November 18, 2020
The latest version of erwin Evolve for enterprise architecture and business process modeling and analysis includes an enhanced integration with erwin Data Modeler so that data models can be ingested by erwin Evolve as part of the wider enterprise architecture. "The more your data and enterprise architecture work together, the more successful your big data, digital transformation and other IT initiatives will be through greater enterprise collaboration," said Martin Owen, erwin's senior vice president of product strategy.
Posted November 18, 2020
NVIDIA unveiled the NVIDIA A100 80GB GPU—the latest innovation powering the NVIDIA HGX AI supercomputing platform—with twice the memory of its predecessor, providing researchers and engineers the performance tools to unlock the next wave of AI and scientific breakthroughs.
Posted November 16, 2020
Solo.io, the modern service connectivity company delivering API infrastructure from the edge to service mesh, announced it has aligned its product suite under the "Gloo" brand name, along with launching the new Gloo Mesh Enterprise. Solo.io's entire set of product offerings will now be organized under the Gloo API Infrastructure Platform.
Posted November 13, 2020
As applications continue to grow in scale and complexity, so does the challenge of supporting them. From geo-distribution, to high-volume transactions and the processing of new data sources and types, modern applications require speed, scalability and flexibility—whether in the ground or the cloud. DBTA recently held a roundtable webinar to educate IT decision-makers and practitioners about the new data management technologies and techniques for speed and scalability.
Posted November 13, 2020
If 2020 taught us anything, it's that no amount of planning can protect organizations from the unforeseen. As IT leaders look at their 2021 plans, what cloud technologies and trends should be top of mind? Where should they place their bets as they consider their budgets and resources? DBTA recently held a webinar featuring Peter Berry, Navisite's CTO of cloud technologies, who reviewed the top five cloud trends to watch in 2021.
Posted November 12, 2020
Splice Machine, a scale-out SQL database with built-in machine learning, is releasing Livewire, its new open source Operational AI platform for industrial IoT use cases.
Posted November 09, 2020
Data modeling has an intimate relationship with abbreviations. Since the creation of the very first data model, there were circumstances where fully worded names for tables or columns were simply too long to implement within one tool or another. Occasionally one runs into an individual who cannot conceive that anyone on the planet would abbreviate something in a different fashion than they do; but more often data modelers tend to enjoy consistency, and when possible, employ rules to support consistent outcomes. Abbreviations are no exception.
Posted November 04, 2020
Videos of presentations from Data Summit Connect Fall 2020, a 3-day series of data management and analytics webinars presented last week by DBTA and Big Data Quarterly, are now available for on-demand viewing on the DBTA YouTube channel. Use this opportunity to take another look, catch up on a session you missed, or share with a colleague!
Posted November 04, 2020
GoodData, a global analytics company, is releasing a new version of the GoodData.UI framework for faster and easier development and delivery of data-driven applications. GoodData also open-sourced its library-making the best practices, principles, and tools available to all application developers.
Posted October 30, 2020
Privitar, a data privacy platform provider, and StreamSets, provider of a DataOps platform, are forming a new partnership and native product integration designed to help organizations accelerate access to data-driven insights. Through the new native integration, users can leverage StreamSets to design and run data pipelines to their data science and data analytics applications, seamlessly applying Privitar's data privacy policies across all of their execution environments.
Posted October 30, 2020
OpenLegacy, a pioneer in digital-driven integration for legacy and core systems, is creating an OEM Partner Program in support of independent software vendors (ISVs) and solution providers. Companies looking to expand their product line and footprint to reach and integrate their enterprise customers' legacy data sources will benefit the most from this program.
Posted October 28, 2020
MemSQL, The Database of Now for operational analytics and cloud-native applications, is rebranding to SingleStore, befitting the organization's expansive value proposition. SingleStore provides a converged data platform for transactions and analytics, which organizations can use to process and analyze their most important time-critical application workloads.
Posted October 27, 2020
Comprehending natural language text with its first-hand challenges of ambiguity, synonymity, and co-reference has been a long-standing problem in natural language processing. Transfer learning uses some of the models that have been pre-trained on terabytes of data and fine-tunes them based on the problem at hand. It's the new way to efficiently implement machine learning solutions without spending months on data cleaning pipeline. Jayeeta Putatunda, senior data scientist, Indellient US Inc., discussed how to implement language model BERT during his Data Summit Connect Fall 2020 session, "The Power of Transfer Learning in NLP using BERT."
Posted October 22, 2020
Qlik is acquiring Blendr.io, an easy-to-use, scalable, and secure embedded integration and automation platform (iPaaS), expanding Qlik's ability to deliver active intelligence through its end-to-end data integration and cloud analytics platform. Blendr.io's technology enables Qlik to expand the existing breadth of its data integration and data analytics offerings to also include automation to trigger actions and enhance augmented analytics capabilities.
Posted October 22, 2020
Hasura, the data access infrastructure company, is adding GraphQL support for MySQL and early access support for SQL Server to its existing support for PostgreSQL, enabling the company to support three of the most popular database technologies.
Posted October 21, 2020
Data Summit Connect Fall began on October 20 with a keynote from Doug Laney, data and analytics innovation fellow, West Monroe and author of "Infonomics," who discussed the new economics of information. He covered the issues of information ownership, rights, and privileges, along with external data opportunities and challenges, and his set of generally accepted information principles culled from other asset management disciplines.
Posted October 20, 2020
OpenStack Foundation has announced it will be changing its corporate name to Open Infrastructure Foundation (OIF) starting in 2021, reflecting the organization's broader commitment and valuable expertise in driving innovations in open source. OIF's goal is to build an open infrastructure for the next decade by solving infrastructure issues raised by continuing demands in new markets.
Posted October 20, 2020
Data Summit Connect Fall 2020, a series of data management and analytics webinars presented by DBTA and Big Data Quarterly, is going on this week. Whether your interests lie in the technical possibilities and challenges of new and emerging technologies or using the wealth of data your company is collecting for business intelligence, analytics, and other business strategies, Data Summit Connect Fall 2020 has something for you!
Posted October 20, 2020
CompassRed, a data analytics and artificial intelligence company, will use its $2 million grant from the CARES Act to fund a Data Innovation Lab to accelerate the use of data insights and intelligence to address COVID-related issues in the Mid-Atlantic region.
Posted October 16, 2020
DataOps has emerged as an agile methodology to improve the speed and accuracy of analytics through new data management practices and processes—from data quality and integration to model deployment and management. Chris Bergh, CEO and head chef, DataKitchen, will lead a workshop on DataOps titled, "Getting Started With DataOps: Orchestrating The Three Pipelines" at the Data Summit Connect Fall 2020 conference.
Posted October 13, 2020
Alation, a provider of enterprise data intelligence solutions, has formed a partnership with Dataiku, an enterprise AI and machine learning platform, to ensure that sensitive data used to create AI and machine-learning models is properly classified and governed. With the Alation and Dataiku integration, data scientists have immediate access to curated data ingested directly into Dataiku.
Posted October 13, 2020
This may be the era of the data-driven enterprise, but only a handful of organizations report they are ready for it. There is a growing volume of "dark data" that remains obscure to IT managers and decision makers. This period unfolding before us will be driven by several technology initiatives, from 5G wireless and IoT to AI.
Posted October 08, 2020
Both data warehouses and data lakes offer robust options for ensuring that data is well-managed and prepped for today's analytics requirements. However, the two environments have distinctly different roles, and data managers need to understand how to leverage the strengths of each to make the most of the data feeding into analytics systems.
Posted October 08, 2020
"It's yours, free. Just pay shipping and handling!" claims the infomercial host. We've all seen them, the advertisements for home maintenance, gardening equipment, or even cooking products that promise the world but end up costing a lot of money. Whether it's through shipping fees or the need to sign up for long contracts that are hard to break, unfortunately, most deals that seem "too good to be true" usually are not what they seem.
Posted October 08, 2020
Rocket Software is in the midst of conducting several product alpha and beta programs to implement user feedback into new and existing products. The fall of 2020 is about Rocket engaging with its MultiValue user base, during these programs customers can provide critical feedback on features they need to modernize and optimize their applications, ranging from API development to helping define metrics that they want to monitor for better performance.
Posted October 08, 2020
Amazon Web Services (AWS) is releasing Amazon Timestream, a new time series database for IoT and operational applications that can scale to process trillions of time series events per day. According to AWS, Amazon Timestream saves customers effort and expense by keeping recent data in-memory and moving historical data to a cost-optimized storage tier based upon user-defined policies, while its query processing gives customers the ability to access and combine recent and historical data transparently across tiers with a single query, without needing to specify explicitly in the query whether the data resides in the in-memory or cost-optimized tier.
Posted October 05, 2020
IT Glue, a Kaseya company and industry standard for IT documentation, is introducing several new product features and integrations, focusing on providing managed service providers (MSPs) and internal IT teams with easier access and greater functionality with their most critical asset - IT documentation.
Posted October 02, 2020
Precisely, a global leader in data integrity, announced the availability of its datasets on the Snowflake Data Marketplace, providing users with a place within the Snowflake Data Marketplace to enrich their own data with Precisely's unique location and consumer insights, for additional context to drive confident business decision-making.
Posted October 01, 2020
Pure Storage, an IT pioneer that delivers storage as-a-service in a multi-cloud world, is introducing a spectrum of enhancements to its hybrid cloud solutions across the VMware portfolio, enabling any enterprise to focus on innovation rather than infrastructure. New solution enhancements from Pure help enterprises maximize their VMware investments, delivering agility and efficiency for modern applications across on-premises and cloud environments. Customers will have improved availability of their data services, enabling them to meet demanding Service-level Agreements (SLAs).
Posted September 29, 2020
Green House Data, a provider of digital transformation consulting and managed IT services, is rebranding to unify its recently acquired companies and nine locations throughout North America under one brand name—Lunavi. The new name reflects the company's growth and ability to provide end-to-end solutions that span from migrating legacy infrastructure to the cloud to enabling cloud native, modern application development, according to the vendor.
Posted September 29, 2020
Ardoq, a software company reinventing enterprise architecture, is launching Scenarios, enabling organizations to model multiple future states in a scalable, data-driven, and collaborative way. Ardoq's Scenarios has taken the lessons learned from collaborative code development and applied them to a data-driven approach for modeling enterprise-wide change.
Posted September 28, 2020
TIBCO Software Inc., a provider of enterprise data solutions, is focusing on disrupting the analytics space with a series of platform releases, including TIBCO Hyperconverged Analytics, providing immersive, smart, and real-time analytics to data-driven businesses. The company also unveiled TIBCO Spotfire 11 and TIBCO Cloud Data Streams, accelerating insights and actions for businesses.
Posted September 24, 2020
Eplexity, a cloud managed services company and Amazon Web Services (AWS) Premier Partner, is collaborating with CloudCheckr, a public cloud governance platform, to provide a comprehensive cloud governance program that enables companies to quickly save on Amazon Web Services (AWS) infrastructure costs.
Posted September 23, 2020
Matillion, the leading provider of data transformation for cloud data warehouses (CDWs),is releasing its free data ingestion tool, Matillion Data Loader, within Snowflake, the cloud data platform, through Snowflake's Partner Connect. Matillion Data Loader enables companies to easily load data into Snowflake from popular data sources including Microsoft SQL Server, Salesforce, Google Analytics, Marketo, and more.
Posted September 23, 2020
NetApp, a global cloud-led, data-centric software company, is releasing Elastigroup for Microsoft Azure Spot Virtual Machines (VMs) after acquiring Spot, a cloud compute management and cost optimization company whose suite of tools includes Elastigroup. Elastigroup from Spot by NetApp simplifies and automates compute infrastructure provisioning and autoscaling while allowing cloud consumers to reduce their compute spending by up to 90%, along with providing high availability, according to the vendor.
Posted September 23, 2020
WANdisco, the LiveData company, is releasing LiveData Migrator, an automated, self-service solution that democratizes cloud data migration at any scale by enabling companies to easily start migrating Hadoop data from on-premises to Amazon Web Services (AWS) within minutes.
Posted September 17, 2020
Instaclustr, delivering reliability at scale through fully managed open source data technologies, is releasing Instaclustr for Redis, enabling low latency in-memory caching, messaging, and stream processing. Redis joins Apache Kafka, Apache Kafka Connect, Apache Cassandra, and Elasticsearch as another highly flexible open source data technology, expertly managed, and supported by Instaclustr.
Posted September 16, 2020
AtScale, a provider of intelligent data virtualization solutions, is adding new datasets and business intelligence (BI) dashboards to the AtScale Data Insights Marketplace, empowering users with up-to-date data on COVID-19. The new datasets include points-of-interest (POI) data, demographic data, campaign finance data, in addition to public economic indicators, and other geospatial data.
Posted September 15, 2020
Aerospike Inc., a provider of next-generation, real-time NoSQL data solutions, is making upgrades and expansions to the Aerospike Connect product line, allowing Aerospike Database 5 to be even more plug-compatible with the enterprise data fabric that powers modern real-time data-driven applications.
Posted September 15, 2020
Never have charts and graphs been more prominent in the collective public consciousness. The increased focus on data-driven insights has, just as so much in life, been both positive and negative.
Posted September 14, 2020
Infoworks.io, a provider of Enterprise Data Operations and Orchestration (EDO2) systems, is making a significant update to its flagship offering, Infoworks DataFoundry, extending data source connectivity to support over 190 different source types. Infoworks customers can now use DataFoundry's enterprise-grade automation to onboard data quickly and easily into the cloud from practically any data source.
Posted September 14, 2020
No matter where you live in the world, your life has been impacted by the technology and innovation conceived of in Silicon Valley. The wealth that has been created has been enormous. A popular British newspaper once published an article suggesting that if Silicon Valley were a country, it would be among the richest on Earth.
Posted September 11, 2020
When talking about data science, most people feel as if they are in one of two camps as far as data size. The first is really small data—hundreds of megabytes to a few gigabytes. The second is gigabytes to terabytes. Notice I didn't say "big data," nor did I say "petabytes." The source datasets may start at petabyte-scale, but keep in mind that data is often very raw, and most of it is ignored.
Posted September 11, 2020