Search for:
What to Expect in AI Data Governance: 2025 Predictions


In 2025, preventing risks from both cyber criminals and AI use will be top mandates for most CIOs. Ransomware in particular continues to vex enterprises, and unstructured data is a vast, largely unprotected asset. AI solutions have moved from experimental to mainstream, with all the major tech companies and cloud providers making significant investments in […]

The post What to Expect in AI Data Governance: 2025 Predictions appeared first on DATAVERSITY.


Read More
Author: Krishna Subramanian

User-Friendly External Smartblobs Using a Shadow Directory

I am very excited about the HCL Informix® 15 external smartblob feature.

If you are not familiar with them, external smartblobs allow the user to store actual Binary Large Object (blob) and Character Large Object (clob) data external to the database. Metadata about that external storage is maintained by the database.

Notes: This article does NOT discuss details of the smartblobs feature itself, but rather proposes a solution to make the functionality more user-friendly. For details on feature behavior, setup, and new functions, see the documentation.

At the writing of this blog, v15.0 does not have the ifx_lo_path function defined, as required below.  This has been reported to engineering.  The workaround is to create it yourself with the following command:

create dba function ifx_lo_path(blob)
  returns lvarchar
  external name '(sq_lo_path)'
  language C;

This article also does not discuss details of client programming required to INSERT blobs and clobs into the database.

The external smartblob feature was built for two main reasons:

1. Backup size

Storing blobs in the database itself can cause the database to become extremely large. As such, performing backups on the database takes an inordinate amount of time, and 0 level backups can be impossible. Offloading the actual blob contents to an external file system can lessen the HCL Informix backup burden by putting the blob data somewhere else. The database still governs the storage of, and access to, the blob, but the physical blob is housed elsewhere/externally.

2. Easy access to blobs

Users would like easy access to blob data, with familiar tools, without having to go through the database. 

Using External Smartblobs in HCL Informix 15

HCL Informix 15 introduces external smartblobs. When you define an external smartblob space, you specify the external directory location (outside the database) where you would like the actual blob data to be stored. Then you assign blob column(s) to that external smartblob space when you CREATE TABLE. When a row is INSERTed, HCL Informix stores the blob data in the defined directory using an internal identifier for the filename.

Here’s an example of a customer forms table: custforms (denormalized and hardcoded for simplicity). My external sbspace directory is /home/informix/blog/resources/esbsp_dir1.

CREATE TABLE custforms(formid SERIAL, company CHAR(20), year INT, lname CHAR(20), 
formname CHAR(50), form CLOB) PUT form IN (esbsp);

Here, I INSERT a 2023 TaxForm123 document from a Java program for a woman named Sanchez, who works for Actian:

try(PreparedStatement p = c.prepareStatement("INSERT INTO custforms 
(company, year, lname, formname, form) values(?,?,?,?,?)");

FileInputStream is = new FileInputStream("file.xml")) {
p.setString(1, "Actian");
p.setString(2, "2023");
p.setString(3, "Sanchez");
p.setString(4, "TaxForm123");
p.setBinaryStream(5, is);
p.executeUpdate();
}

After I INSERT this row, my external directory and file would look like this:

[informix@schma01-rhvm03 resources]$ pwd
/home/informix/blog/resources
[informix@schma01-rhvm03 resources]$ ls -l esbsp*
-rw-rw---- 1 informix informix 10240000 Oct 17 13:22 esbsp_chunk1

esbsp_dir1:
total 0
drwxrwx--- 2 informix informix 41 Oct 17 13:19 IFMXSB0
[informix@schma01-rhvm03 resources]$ ls esbsp_dir1/IFMXSB0
LO[2,2,1(0x102),1729188125]

Where LO[2,2,1(0x102),1729188125]is an actual file that contains the data that I could access directly. The problem is that if I want to directly access this file for Ms. Sanchez, I would first have to figure out that this file belongs to her and is the tax document I want. It’s very cryptic!

A User-Friendly Smartblob Solution

When talking to Informix customers, they love the new external smartblobs feature but wish it could be a little more user-friendly.

As in the above example, instead of putting Sanchez’s 2023 TaxForm123 into a general directory called IFMXSB0 in a file called LO[2,2,1(0x102),1729188125, which together are meaningless to an end-user, wouldn’t it be nice if the file was located in an intuitive place like /home/forms/Actian/2024/TaxForm123/Sanchez.xml or something similar…something meaningful…how YOU want it organized?

Having HCL Informix automatically do this is a little easier said than done, primarily because the database would not intuitively know how any one customer would want to organize their blobs. What exact directory substructure? From what column or columns do I form the file names? What order? All use cases would be different.

Leveraging a User-Friendly Shadow Directory

The following solution shows how you can create your own user-friendly logical locations for your external smartblobs by automatically maintaining a lightweight shadow directory structure to correspond to actual storage locations. The solution uses a very simple system of triggers and stored procedures to do this.

Note: Examples here are shown on Linux, but other UNIX flavors should work also.

How to Set Up in 4 Steps

For each smartblob column in question

STEP 1: Decide how you want to organize access to your files.

Decide what you want the base of your shadow directory to be and create it. In my case for this blog, it is: /home/informix/blog/resources/user-friendly. You could probably implement this solution without a set base directory (as seen in the examples), but that may not be a good idea because users would unknowingly start creating directories everywhere.

STEP 2: Create a create_link stored procedure and corresponding trigger for INSERTs.

This procedure makes sure that the desired data-driven subdirectory structure exists from the base (mkdir -p), then forms a user-friendly logical link to the Informix smartblob file.    You must pass all the columns to this procedure from which you want to form the directory structure and filename from the trigger.

CREATE PROCEDURE

CREATE PROCEDURE create_link (p_formid INT, p_company CHAR(20), p_year INT,
p_lname CHAR(20), p_formname CHAR(50))
DEFINE v_oscommand CHAR(500);
DEFINE v_custlinkname CHAR(500);
DEFINE v_ifmxname CHAR(500);
DEFINE v_basedir CHAR(100);
-- set the base directory
LET v_basedir = '/home/informix/blog/resources/user-friendly';
-- make sure directory tree exists
LET v_oscommand = 'mkdir -p ' || TRIM(v_basedir) || '/' || TRIM(p_company) || '/' || 
TO_CHAR(p_year);
SYSTEM v_oscommand; 

-- form full link name 
LET v_custlinkname = TRIM(v_basedir) || '/' || TRIM(p_company) || '/' || TO_CHAR(p_year) 
|| '/' || TRIM(p_lname) || '.' || TRIM(p_formname) || '.' || TO_CHAR(p_formid);

-- get the actual location 
SELECT IFX_LO_PATH(form::LVARCHAR) INTO v_ifmxname FROM custforms WHERE formid = p_formid; 

-- create the os link 
LET v_oscommand = 'ln -s -f ' || '''' || TRIM(v_ifmxname) || '''' || ' ' || v_custlinkname; 
SYSTEM v_oscommand;

END PROCEDURE

CREATE TRIGGER

CREATE TRIGGER ins_tr INSERT ON custforms REFERENCING new AS post
FOR EACH ROW(EXECUTE PROCEDURE create_link (post.formid, post.company,
post.year, post.lname, post.formname));

STEP 3: Create a delete_link stored procedure and corresponding trigger for DELETEs.

This procedure will delete the shadow directory link if the row is deleted.

CREATE PROCEDURE

CREATE PROCEDURE delete_link (p_formid INT, p_company CHAR(20), p_year INT,
p_lname CHAR(20), p_formname CHAR(50))
DEFINE v_oscommand CHAR(500);
DEFINE v_custlinkname CHAR(500); 
DEFINE v_basedir CHAR(100);
-- set the base directory
LET v_basedir = '/home/informix/blog/resources/user-friendly';
-- form full link name
LET v_custlinkname = TRIM(v_basedir) || '/' ||
TRIM(p_company) || '/' || TO_CHAR(p_year) || '/' || TRIM(p_lname) || '.'
|| TRIM(p_formname) || '.' || TO_CHAR(p_formid);
-- remove the link
LET v_oscommand = 'rm -f -d ' || v_custlinkname;
SYSTEM v_oscommand;

END PROCEDURE

CREATE TRIGGER

CREATE TRIGGER del_tr DELETE ON custforms REFERENCING old AS pre FOR EACH ROW
(EXECUTE PROCEDURE delete_link (pre.formid, pre.company, pre.year, pre.lname, pre.formname));

STEP 4: Create a change_link stored procedure and corresponding trigger for UPDATEs, if desired.   In my example, Ms. Sanchez might get married to Mr. Simon and an UPDATE to her last name in the database occurs. I may then want to change all my user-friendly names from Sanchez to Simon.  This procedure deletes the old link and creates a new one.

Notice the update trigger only must fire on the columns that form your directory structure and filenames.

CREATE PROCEDURE

CREATE PROCEDURE change_link (p_formid INT, p_pre_company CHAR(20), 
p_pre_year INT, p_pre_lname CHAR(20), p_pre_formname CHAR(50), p_post_company CHAR(20), 
p_post_year INT, p_post_lname CHAR(20), p_post_formname CHAR(50))

DEFINE v_oscommand CHAR(500);
DEFINE v_custlinkname CHAR(500);
DEFINE v_ifmxname CHAR(500);
DEFINE v_basedir CHAR(100);
-- set the base directory
LET v_basedir = '/home/informix/blog/resources/user-friendly';

-- get rid of old

-- form old full link name
LET v_custlinkname = TRIM(v_basedir) || '/' || TRIM(p_pre_company) || '/' || 
TO_CHAR(p_pre_year) || '/' || TRIM(p_pre_lname) || '.' || TRIM(p_pre_formname) || '.' 
|| TO_CHAR(p_formid) ;

-- remove the link and empty directories
LET v_oscommand = 'rm -f -d ' || v_custlinkname;
SYSTEM v_oscommand;

-- form the new
-- make sure directory tree exists
LET v_oscommand = 'mkdir -p ' || TRIM(v_basedir) || '/' || TRIM(p_post_company) || '/' || 
TO_CHAR(p_post_year);
SYSTEM v_oscommand;

-- form full link name
LET v_custlinkname = TRIM(v_basedir) || '/' || TRIM(p_post_company) || '/' || 
TO_CHAR(p_post_year) || '/' || TRIM(p_post_lname) || '.' || TRIM(p_post_formname) 
|| '.' || TO_CHAR(p_formid) ;

-- get the actual location
-- this is the same as before as id has not changed
SELECT IFX_LO_PATH(form::LVARCHAR) INTO v_ifmxname FROM custforms WHERE formid = p_formid;

-- create the os link
LET v_oscommand = 'ln -s -f ' || '''' || TRIM(v_ifmxname) || '''' || ' ' || v_custlinkname;
SYSTEM v_oscommand;

END PROCEDURE

CREATE TRIGGER

CREATE TRIGGER upd_tr UPDATE OF formid, company, year, lname, formname ON custforms
REFERENCING OLD AS pre NEW as post

FOR EACH ROW(EXECUTE PROCEDURE change_link (pre.formid, pre.company, pre.year, pre.lname, 
pre.formname, post.company, post.year, post.lname, post.formname));

Results Example

Back to our example.

With this infrastructure in place, now in addition to the Informix-named file being in place, I would have these user-friendly links on my file system that I can easily locate and identify.

INSERT

[informix@schma01-rhvm03 2023]$ pwd
/home/informix/blog/resources/user-friendly/Actian/2023
[informix@schma01-rhvm03 2023]
$ ls Sanchez.TaxForm123.2

If I do an ls -l, you will see that it is a link to the Informix blob file.

[informix@schma01-rhvm03 2023]$ ls -l
total 0
lrwxrwxrwx 1 informix informix 76 Oct 17 14:20 Sanchez.TaxForm123.2 -> 
/home/informix/blog/resources/esbsp_dir1/IFMXSB0/LO[2,2,1(0x102),1729188126]

UPDATE

If I then update her last name with UPDATE custforms SET lname = ‘Simon’ where formid=2,my file system now looks like this:

[informix@schma01-rhvm03 2023]$ ls -l
lrwxrwxrwx 1 informix informix 76 Oct 17 14:25 Simon.TaxForm123.2 -> 
/home/informix/blog/resources/esbsp_dir1/IFMXSB0/LO[2,2,1(0x102),1729188126]

DELETE

If I then go and DELETE this form with DELETE FROM custforms where formid=2, my directory structure looks like this:

[informix@schma01-rhvm03 2023]$ pwd
/home/informix/blog/resources/user-friendly/Actian/2023
[informix@schma01-rhvm03 2023]$ ls
[informix@schma01-rhvm03 2023]$

We Welcome Your Feedback

Please enjoy the new HCL Informix15 external smartblob feature.

I hope this idea can make external smartblobs easier for you to use. If you have any feedback on the idea, especially on enhancements or experience in production, please feel free to contact me at mary.schulte@hcl-software.com. I look forward to hearing from you!

Find out more about the launch of HCL Informix 15.

Notes

1. Shadow directory permissions. In creating this example, I did not explore directory and file permissions, but rather just used general permissions settings on my sandbox server. Likely, you will want to control permissions to avoid some of the anomalies I discuss below.

2. Manual blob file delete. With external smartblobs, if permissions are not controlled, it is possible that a user might somehow delete the physical smartblob file itself from its directory. HCL Informix, itself, cannot control this from happening. In the event it does happen, HCL Informix does NOT delete the corresponding row; the blob file will just be missing. There may be aspects to links that can automatically handle this, but I have not investigated them for this blog.

3. Link deletion in the shadow directory. If permissions are not controlled, it is possible that a user might delete a logical link formed by this infrastructure. This solution does not detect this. If this is an issue, I would suggest a periodic maintenance job that cross references the shadow directory links to blob files to detect missing links. For those blobs with missing links, write a database program to look up the row’s location with the IFX_LO_PATH function, and reform the missing link.

4. Unique identifiers. I highly recommend using unique identifiers in this solution. In this simple example, I used formid. You don’t want to clutter things up, of course, but depending on how you structure your shadow directories and filenames, you may need to include more unique identifiers to avoid directory and link names duplication.

5. Empty directories. I did not investigate if there are options to rm in the delete stored procedure to clean up empty directories that might remain if a last item is deleted.

6. Production overhead. It is known that excessive triggers and stored procedures can add overhead to a production environment. For this blog, it is assumed that OLTP activity on blobs is not excessive, therefore production overhead should not be an issue. This being said, this solution has NOT been tested at scale.

7. NULL values. Make sure to consider the presence and impact of NULL values in columns used in this solution. For simplicity, I did not handle them here.

Informix is a trademark of IBM Corporation in at least one jurisdiction and is used under license.

 

The post User-Friendly External Smartblobs Using a Shadow Directory appeared first on Actian.


Read More
Author: Mary Schulte

AI Predictions for 2025: Embracing the Future of Human and Machine Collaboration


Predictions are funny things. They often seem like a bold gamble, almost like trying to peer into the future with the confidence we inherently lack as humans. Technology’s rapid advancement surprises even the most seasoned experts, especially when it progresses exponentially, as it often does. As physicist Albert A. Bartlett famously said, “The greatest shortcoming […]

The post AI Predictions for 2025: Embracing the Future of Human and Machine Collaboration appeared first on DATAVERSITY.


Read More
Author: Philip Miller

Focusing on Data Privacy: Building Trust and Strengthening Security


In today’s digital age, managing and minimizing data collection is essential for maintaining business security. Prioritizing data privacy helps organizations ensure they only gather necessary information, reducing the risk of data breaches and misuse. This approach addresses potential vulnerabilities at their source, mitigating the impact of breaches and avoiding regulatory penalties as scrutiny over privacy […]

The post Focusing on Data Privacy: Building Trust and Strengthening Security appeared first on DATAVERSITY.


Read More
Author: Dorababu Nadella

Transforming Marketing Data into Business Growth: Key Insights and Strategies


Marketing leaders and data professionals often grapple with a familiar challenge: how to transform marketing data into tangible business growth. During a recent episode of The Lights on Data Show, I had the privilege of speaking with Kasper Bossen-Rasmussen, founder and CEO of Accutics, about this very topic. Together, we explored key takeaways for addressing […]

The post Transforming Marketing Data into Business Growth: Key Insights and Strategies appeared first on LightsOnData.


Read More
Author: George Firican

Why the Growing Adoption of IoT Demands Seamless Integration of IT and OT


Over the past year, cyberattacks on cyber-physical systems (CPS) have cost organizations around the world at least $500,000, highlighting the growing financial and operational risks of compromised security. As artificial intelligence (AI) continues to emerge as a key driver in nearly every sector, the need for trustworthy, secure data becomes even more crucial. To address these challenges, […]

The post Why the Growing Adoption of IoT Demands Seamless Integration of IT and OT appeared first on DATAVERSITY.


Read More
Author: Julian Durand

Alternatives to Azure Document Intelligence Studio: Exploring Powerful Document Analysis Tools


Document Intelligence Studio is a data extraction tool that can pull unstructured data from diverse documents, including invoices, contracts, bank statements, pay stubs, and health insurance cards. The cloud-based tool from Microsoft Azure comes with several prebuilt models designed to extract data from popular document types. However, you can also use labeled datasets to train…
Read more

The post Alternatives to Azure Document Intelligence Studio: Exploring Powerful Document Analysis Tools appeared first on Seattle Data Guy.


Read More
Author: research@theseattledataguy.com

From Input to Insight: How Quality Data Drives AI and Automation


More and more enterprises are looking to automation and AI to deliver new efficiencies and give their organizations an edge in the market. Data is the engine that powers both automation and AI. But data must be clean and user-friendly for these systems to work effectively and deliver on their promise.  Lots of organizations are […]

The post From Input to Insight: How Quality Data Drives AI and Automation appeared first on DATAVERSITY.


Read More
Author: Amol Dalvi

Data Monetization: The Holy Grail or the Road to Ruin?


Unlocking the value of data is a key focus for business leaders, especially the CIO. While in its simplest form, data can lead to better insights and decision-making, companies are pursuing an entirely different and more advanced agenda: the holy grail of data monetization. This concept involves aggregating a variety of both structured and unstructured […]

The post Data Monetization: The Holy Grail or the Road to Ruin? appeared first on DATAVERSITY.


Read More
Author: Tony Klimas

Modern Data Archiving: Managing Explosive Unstructured Data Growth


As unstructured data creation rates have soared, the timeframe for active use of data has shrunk due to edge computing, IoT systems, machine-generated data, and, let’s not forget, GenAI. The data use period today has largely been reduced to around 30 to 90 days before the flood of new data appearing makes the existing data […]

The post Modern Data Archiving: Managing Explosive Unstructured Data Growth appeared first on DATAVERSITY.


Read More
Author: Steve Leeper

Delivering Personalized Recommendations Without Sacrificing User Privacy


In today’s fast-paced digital landscape, we all love a little bit of personalization. Whether it’s Netflix suggesting our next binge-worthy show or Spotify curating our playlists, these tailored experiences make us feel understood and valued. But with growing concerns around user privacy, how can companies achieve this level of personalization without compromising our personal data? […]

The post Delivering Personalized Recommendations Without Sacrificing User Privacy appeared first on DATAVERSITY.


Read More
Author: Ganapathy Subramanian Ramachandran

Beyond Ownership: Scaling AI with Optimized First-Party Data


Brands, publishers, MarTech vendors, and beyond recently gathered in NYC for Advertising Week and swapped ideas on the future of marketing and advertising. The overarching message from many brands was one we’ve heard before: First-party data is like gold, especially for personalization. But it takes more than “owning” the data to make it valuable. Scale and accuracy […]

The post Beyond Ownership: Scaling AI with Optimized First-Party Data appeared first on DATAVERSITY.


Read More
Author: Tara DeZao

Accelerating Innovation: Data Discovery in Manufacturing

The manufacturing industry is in the midst of a digital revolution. You’ve probably heard these buzzwords: Industry 4.0, IoT, AI, and machine learning– all terms that promise to revolutionize everything from assembly lines to customer service. Embracing this digital transformation is key in improving your competitive advantage, but new technology doesn’t come without its own challenges. Each new piece of technology needs one thing to deliver innovation: data.

Data is the fuel powering your tech engines. Without the ability to understand where your data is, whether it’s trustworthy, or who owns the datasets, even the most powerful tools can overcomplicate and confuse the best data teams. That’s where modern data discovery solutions come in. They’re like the backstage crew making sure everything runs smoothly– connecting systems, tidying up the data mess, and making sure everyone has exactly what they need, when they need it. That means faster insights, streamlined operations, and a lower total cost of ownership (TCO). In other words, data access is the key to staying ahead in today’s fast-paced, highly competitive, increasingly sensitive manufacturing market. 

The Problem

Data from all aspects of your business is siloed– whether it’s coming from sensors, legacy systems, cloud applications, suppliers or customers– trying to piece it all together is daunting, time-consuming, and just plain hard. Traditional methods are slow, cumbersome, and definitely not built for today’s needs. This fragmented approach not only slows down decision-making, but keeps you from tapping into valuable insights that could drive innovation. And in a market where speed is everything, that’s a recipe for falling behind. 

So the big question is: how can you unlock the true potential of your data?

The Solution

So how do you make data intelligence into a streamlined, efficient process? The answer lies in modern data discovery solutions– the unsung catalyst of a digital transformation motion. Rather than simply integrating data sources, data discovery solutions excel in metadata management, offering complete visibility into your company’s data ecosystem. They enable users– regardless of skill level– to locate where data resides and assess the quality and relevance of the information. By providing this detailed understanding of data context and lineage, organizations can confidently leverage accurate, trustworthy datasets, paving the way for informed decision-making and innovation, 

Key Components

Easy-to-Connect Data Sources for Metadata Management

 One of the biggest hurdles in data integration is connecting to a variety of data sources, including legacy systems, cloud applications, and IoT devices. Modern data discovery tools like Zeenea offer easy connectivity, allowing you to extract metadata from various sources seamlessly. This unified view eliminates silos and enables faster, more informed decision-making across the organization.

Advanced Metadata Management

Metadata is the backbone of effective data discovery. Advanced metadata management capabilities ensure that data is well-organized, tagged, and easily searchable. This provides a clear context for data assets, helping you understand the origin, quality, and relevance of your data. This means better data search and discoverability.

Data Discovery Knowledge Graph

A data discovery knowledge graph serves as an intelligent map of your metadata, illustrating the intricate relationship and connections between data assets. It provides users with a comprehensive view of how data points are linked across systems, offering a clear picture of data lineage– from origin to current state. The visibility into the data journey is invaluable in manufacturing, where understanding the flow of information between production data, supply chain metrics, and customer feedback is critical. By tracing the lineage of data, you can quickly assess its accuracy, relevance, and context, leading to more precise insights and informed decision-making.

Quick Access to Quality Data Through Data Marketplace

A data marketplace provides a centralized hub where you can easily search, discover, and access high-quality data. This self-service model empowers your teams to find the information they need without relying on IT, accelerating time to insight. The result? Faster product development cycles, improved process efficiency, and enhanced decision-making capabilities.

User-Friendly Interface With Natural Language Search

Modern data discovery platforms prioritize user experience with intuitive, user-friendly interfaces. Features like natural language search allow users to query data using everyday language, making it easier for non-technical users to find what they need. This democratizes access to data across the organization, fostering a culture of data-driven decision-making.

Low Total Cost of Ownership (TCO)

Traditional metadata management solutions often come with a hefty price tag due to high infrastructure costs and ongoing maintenance. In contrast, modern data discovery tools are designed to minimize TCO with automated features, cloud-based deployment, and reduced need for manual intervention. This means more efficient operations and a greater return on investment.

Benefits

By leveraging a comprehensive data discovery solution, manufacturers can achieve several key benefits:

Enhanced Innovation

With quick access to quality data, teams can identify trends and insights that drive product development and process optimization.

Faster Time to Market

Automated implementation and seamless data connectivity reduce the time required to gather and analyze data, enabling faster decision-making.

Improved Operational Efficiency

Advanced metadata management and knowledge graphs help streamline data governance, ensuring that users have access to reliable, high-quality data.

Increased Competitiveness

A user-friendly data marketplace democratizes data access, empowering teams to make data-driven decisions and stay ahead of industry trends.

Cost Savings

With low TCO and reduced dependency on manual processes, manufacturers can maximize their resources and allocate budgets towards strategic initiatives.

Data is more than just a resource—it’s a catalyst for innovation. By embracing advanced metadata management and data discovery solutions, you can find, trust, and access data. This not only accelerates time to market but also drives operational efficiency and boosts competitiveness. With powerful features like API-led automation, a data discovery knowledge graph, and an intuitive data marketplace, you’ll be well-equipped to navigate the challenges of Industry 4.0 and beyond.

Call to Action

Ready to accelerate your innovation journey? Explore how Actian Zeenea can transform your manufacturing processes and give you a competitive edge.

Learn more about how our advanced data discovery solutions can help you unlock the full potential of your data. Sign up for a live product demo and Q&A

 

The post Accelerating Innovation: Data Discovery in Manufacturing appeared first on Actian.


Read More
Author: Kasey Nolan

Ask a Data Ethicist: How Can You Learn More About Data and AI Ethics?


It was about this time last year that I pitched the team at DATAVERSITY the idea of this monthly column on data ethics. There’s certainly been no shortage of interesting questions to cover and I’ve enjoyed writing about both the practical and more philosophical aspects of this topic. As we wrap up this year and […]

The post Ask a Data Ethicist: How Can You Learn More About Data and AI Ethics? appeared first on DATAVERSITY.


Read More
Author: Katrina Ingram

The Hidden Infrastructure Crisis: Why CIOs Face a Perfect Storm in IT Talent Management


As organizations navigate the complex landscape of digital transformation, CIOs are confronting an unprecedented crisis that extends far beyond the typical challenges of recruitment and retention. At its core, this crisis represents a fundamental misalignment between traditional IT infrastructure management and modern development practices – a gap that threatens to widen as experienced IT professionals retire […]

The post The Hidden Infrastructure Crisis: Why CIOs Face a Perfect Storm in IT Talent Management appeared first on DATAVERSITY.


Read More
Author: Daniel Clydesdale-Cotter

Navigating the Complex Landscape of Data Sovereignty


In today’s rapidly evolving global landscape, data sovereignty has emerged as a critical challenge for enterprises. Businesses must adapt to an increasingly complex web of requirements as countries around the world tighten data regulations in an effort to ensure compliance and protect against cyberattacks. Data sovereignty regulations significantly impact an organization’s ability to conduct data […]

The post Navigating the Complex Landscape of Data Sovereignty appeared first on DATAVERSITY.


Read More
Author: Mark Cusack

Mind the Gap: Architecting Santa’s List – The Naughty-Nice Database


You never know what’s going to happen when you click on a LinkedIn job posting button. I’m always on the lookout for interesting and impactful projects, and one in particular caught my attention: “Far North Enterprises, a global fabrication and distribution establishment, is looking to modernize a very old data environment.” I clicked the button […]

The post Mind the Gap: Architecting Santa’s List – The Naughty-Nice Database appeared first on DATAVERSITY.


Read More
Author: Mark Cooper

From Silos to Synergy: Data Discovery for Manufacturing

Introduction

There is an urgent reality that many manufacturing leaders are facing, and that’s data silos. Valuable information remains locked within departmental systems, hindering your ability to make strategic, well-informed decisions. A data catalog and enterprise data marketplace solution provides a comprehensive, integrated view of your organization’s data, breaking down silos and enabling true collaboration. 

The Problem: Data Silos Impede Visibility

In your organization, each department maintains its own critical datasets– finance compiles detailed financial reports, sales leverages CRM data, marketing analyzes campaign performance, and operations tracks supply chain metrics. But here’s the challenge: how confident are you that you even know what data is available, who owns it, or if it’s quality?

The issue goes beyond traditional data silos. It’s not just that the data is isolated– it’s that your teams are unaware of what data even exists. This lack of visibility creates a blind spot. Without a clear understanding of your company’s data landscape, you face inefficiencies, inconsistent analysis, and missed opportunities. Departments and up duplicating work, using outdated or unreliable data, and making decisions based on incomplete information.

The absence of a unified approach to data discovery and cataloging means that even if the data is technically accessible, it remains hidden in plain sight, trapped in disparate systems without any context or clarity. Without a comprehensive search engine for your data, your organization will struggle to:

  • Identify data sources: You can’t leverage data if you don’t know it exists. Without visibility into all available datasets, valuable information often remains unused, limiting your ability to make fully informed decisions.
  • Access data quality: Even when you find the data, how do you know it’s accurate and up-to-date? Lack of metadata means you can’t evaluate the quality or relevance of the information, leading to analysis based on faulty data.
  • Understand data ownership: when it’s unclear who owns or manages specific datasets, you waste time tracking down information and validating its source. This confusion slows down projects and introduces unnecessary friction. 

The Solution

Now, imagine the transformative potential if your team could search for and discover all available data across your organization as easily as using a search engine. Implementing a robust metadata management strategy—including data lineage, discovery, and cataloging—bridges the gaps between disparate datasets, enabling you to understand what data exists, its quality, and how it can be used. Instead of chasing down reports or sifting through isolated systems, your teams gain an integrated view of your company’s data assets.

  • Data Lineage provides a clear map of how data flows through your systems, from its origin to its current state. It allows you to trace the journey of your data, ensuring you know where it came from, how it’s been transformed, and if it can be trusted. This transparency is crucial for verifying data quality and making accurate, data-driven decisions.
  • Data Discovery enables teams to quickly search through your company’s data landscape, finding relevant datasets without needing to know the specific source system. It’s like having a powerful search tool that surfaces all available data, complete with context about its quality and ownership, helping your team unlock valuable insights faster.
  • A Comprehensive Data Catalog serves as a central hub for all your metadata, documenting information about the datasets, their context, quality, and relationships. It acts as a single source of truth, making it easy for any team member to understand what data is available, who owns it, and how it can be used effectively.

Revolutionizing Your Operations With Metadata Management

This approach can transform the way each department operates, fostering a culture of informed decision-making and reducing inefficiencies:

  • Finance gains immediate visibility into relevant sales data, customer demand forecasts, and historical trends, allowing for more accurate budgeting and financial planning. With data lineage, your finance team can verify the source and integrity of financial metrics, ensuring compliance and minimizing risks.
  • Sales can easily search for and access up-to-date product data, customer insights, and market analysis, all without needing to navigate complex systems. A comprehensive data catalog simplifies the process of finding the most relevant datasets, enabling your sales team to tailor their pitches and close deals faster.
  • Marketing benefits from an integrated view of customer behavior, campaign performance, and product success. Using data discovery, your marketing team can identify the most impactful campaigns and refine strategies based on real-time feedback, driving greater engagement and ROI.
  • Supply Chain Leaders can trace inventory data back to its origin, gaining full visibility into shipments, supplier performance, and potential disruptions. With data lineage, they understand the data’s history and quality, allowing for proactive adjustments and optimized procurement.
  • Manufacturing Managers have access to a clear, unified view of production data, demand forecasts, and operational metrics. The data catalog offers a streamlined way to integrate insights from across the company, enabling better decision-making in scheduling, resource allocation, and quality management.
  • Operations gains a comprehensive understanding of the entire production workflow, from raw materials to delivery. Data discovery and lineage provide the necessary context for making quick adjustments, ensuring seamless production and minimizing delays.

This strategy isn’t about collecting more data—it’s about creating a clearer, more reliable picture of your entire business. By investing in a data catalog, you turn fragmented insights into a cohesive, navigable map that guides your strategic decisions with clarity and confidence. It’s the difference between flying blind and having a comprehensive navigation system that leads you directly to success.

The Benefits: From Fragmentation to Unified Insight

When you prioritize data intelligence with a catalog as a cornerstone, your organization gains access to a powerful suite of benefits:

  1. Enhanced Decision-Making: With a unified view of all data sources, your team can make well-informed decisions based on real-time insights. Data lineage allows you to trace back the origin of key metrics, ensuring the accuracy and reliability of your analysis.
  2. Improved Collaboration Across Teams: With centralized metadata and clear data relationships, every department has access to the same information, reducing silos and fostering a culture of collaboration.
  3. Greater Efficiency and Reduced Redundancies: By eliminating duplicate efforts and streamlining data access, your teams can focus on strategic initiatives rather than time-consuming data searches.
  4. Proactive Risk Management: Full visibility into data flow and origins enables you to identify potential issues before they escalate, minimizing disruptions and maintaining smooth operations.
  5. Increased Compliance and Data Governance: Data lineage provides a transparent trail for auditing purposes, ensuring your organization meets regulatory requirements and maintains data integrity.

Conclusion

Data silos are more than just an operational inconvenience—they are a barrier to your company’s growth and innovation. By embracing data cataloging, lineage, and governance, you empower your teams to collaborate seamlessly, leverage accurate insights, and make strategic decisions with confidence. It is time to break down the barriers, integrate your metadata, and unlock the full potential of your organization’s data.

Call to Action

Are you ready to eliminate data silos and gain a unified view of your operations? Discover the power of metadata management with our comprehensive platform. Visit our website today to learn more and sign up for a live product demo and Q&A.

The post From Silos to Synergy: Data Discovery for Manufacturing appeared first on Actian.


Read More
Author: Kasey Nolan

5 Data Management Tool and Technology Trends to Watch in 2025


The market surrounding data management tools and technologies is quite mature. After all, the typical business has been making extensive use of data to help streamline its operations and decision-making for years, and many companies have long had data management tools in place. But that doesn’t mean that little is happening in the world of […]

The post 5 Data Management Tool and Technology Trends to Watch in 2025 appeared first on DATAVERSITY.


Read More
Author: Matheus Dellagnelo

How to Foster a Cross-Organizational Approach to Data Initiatives


In today’s business landscape, data reigns supreme. It is the cornerstone of effective decision-making, fuels innovation, and drives organizational success. However, despite its immense potential, many organizations struggle to harness the full power of their data due to a fundamental disconnect between IT and business teams. This division not only impedes progress but also undermines […]

The post How to Foster a Cross-Organizational Approach to Data Initiatives appeared first on DATAVERSITY.


Read More
Author: Abhas Ricky

Data Governance Defying Gravitas
“Defying Gravity,” the show-stopping anthem from the musical “Wicked,” captures the essence of breaking free from conventions and soaring beyond expectations. Just as Elphaba, the protagonist witch from “Wicked,” refuses to be bound by the weight of societal norms, Non-Invasive Data Governance (NIDG) offers organizations a way to defy the gravitas of traditional governance frameworks. […]


Read More
Author: Robert S. Seiner

Through the Looking Glass: What Does Data Quality Mean for Unstructured Data?
I go to data conferences. Frequently. Almost always right here in NYC. We have lots of data conferences here. Over the years, I’ve seen a trend — more and more emphasis on AI.   I’ve taken to asking a question at these conferences: What does data quality mean for unstructured data? This is my version of […]


Read More
Author: Randall Gordon

Data Governance Best Practices: Lessons from Anthem’s Massive Data Breach
In the insurance industry, data governance best practices are not just buzzwords — they’re critical safeguards against potentially catastrophic breaches. The 2015 Anthem Blue Cross Blue Shield data breach serves as a stark reminder of why robust data governance is crucial.  The Breach: A Wake-Up Call  In January 2015, Anthem, one of the largest health […]


Read More
Author: Christine Haskell

Data Insights Ensure Quality Data and Confident Decisions
Every business (large or small) creates and depends upon data. One hundred years ago, businesses looked to leaders and experts to strategize and to create operational goals. Decisions were based on opinion, guesswork, and a complicated mixture of notes and records reflecting historical results that may or may not be relevant to the future.  Today, […]


Read More
Author: Kartik Patel

RSS
YouTube
LinkedIn
Share