Ask a question, get and give advice, share your tips & tricks, expand your network, discuss data topics. This place is for you.
- 196 Topics
- 391 Replies
Enjoy Data People Summit 2022 On Demand
Join the Ataccama Data People Summit 2022 - Virtual Attendees group for our virtual event attendees, and browse recordings from the keynote, panel, and product deep dive sessions. You can catch anything you might have missed, share your thoughts on the sessions, and connect with fellow virtual attendees to continue the conversation.Browse recordings Looking for more? Share a summary of the event with your teamEnjoy a refresher of all the sessions that took place at the Data People Summit, browse the photo gallery, and share access to select keynotes with any colleagues who couldn’t join on-site.Share Sign up for our client-led user groupWe’re thrilled to share that Catherine Yoshida, Manager, Data Governance & Data Architecture at Teranet is chairing virtual Ataccama Client-Led User Group (ACLUG) each quarterly. Join fellow data practitioners with similar interests and a common focus on Ataccama for the kickoff on December 8th.Sign up The latest data quality news from GartnerWe ta
A Leader again! 2022 Gartner Magic Quadrant for Data Quality SolutionsNews
We are thrilled to share the news that Ataccama has been named a Leader in the 2022 Gartner® Magic Quadrant™ for Data Quality Solutions. Access your copy of the complimentary report to learn how data quality technology has changed in the past year, what to look for in data quality software to support your program, what makes Ataccama a Leader in the Magic Quadrant, and more: https://bit.ly/3t5d5Mh
How Ataccama ONE uses Snowpark data frames to automate profiling
Ataccama is known and appreciated by our customers for complex, big deployments where massive amounts of data are processed. We were one of the first to support MapReduce and Spark as native pushdown and enable distributed processing across different cluster nodes.Ataccama’s engine works with the major of enterprise data sources and processes them to provide valuable insight, such as statistics of the data, detection of domains, and data quality.To provide the best experience for our customers that use Snowflake, we developed a tighter integration between our technologies. Specifically, we use the Snowflake data processing capabilities in a pushdown manner which means that data is processed directly on Snowflake.The key advantages of this integration are:Lower data transfer costs for the client. Snowflake’s computational capabilities deliver results faster because of reduced IO operations. Customers don’t incur additional costs to maintain external systems for Big Data. Improved securi
Comparing values from two tables
Hi,Situation: I wrote a(n) SQL catalog item joining multiple tables and in Ataccama I'm using rules to compare the values if they are the same. LONG = LONG, check if the nr_loan = the same a.nr_loan, etc.My question is if there's an easier way to see if two tables are identical. I'm doing this to check the integrity.I couldn't find anything relatable asked in the community.
Remote executor and datamart
From the ataccama one url, click environment settings, create environment we have the below properties executor endpointadmin centerexecution mode then we have the option to add a datamart, with a jdbc string. What exactly is the executor endpoint, admin center and execution mode?is the datamart a place(database) to store something from the executor(server)?Please provide some links which explains this.
What is Global runtime configuration and why we need it
Global runtime configuration (GRC) is the answer to the following questions:How to upload DQ results from my monitoring project into a database/other data source instead of CSV file? How to send my DQ results into PowerBI to create some nice reports? How to connect to my data source from a VCI if this data source is not supported from the Catalog?These three questions are the most frequent use cases for using the Global runtime configuration. It allows us to define additional connections to databases, servers (e.g. MiniO or S3), APIs etc. and perform different kinds of operations with it - like save our DQ results in a post-processing component to a database or upload data from additional data sources using the VCI (Virtual Catalog Item).How to create a Global runtime configurationThe GRC is attached as a parameter plugin.dqc-support.ataccama.one.dqc.runtime-config.default in the DPM application.properties (or using the Configuration Service). What is very important, the GRC is an XML
OAuth 2.0 via Keycloak for API RequestsBest Practice
OAuth 2.0 is an authorization framework that allows third-party applications to access resources in a secure manner. Keycloak is an open-source identity and access management solution that provides OAuth 2.0 authorization services. This article will provide steps on how to use OAuth 2.0 via Keycloak for API requests.First, create a client in Keycloak. This client will contain the necessary information needed to make API requests. In the Keycloak admin console, navigate to the "Clients" section and click "Create". Fill in the necessary information, such as the client ID and save the Client.Creating a Client in Keycloak Once saved, change the Access Type to "Confidential". Additionally, modify the Valid Redirect URL, to the base url of the application and an asterisk. For example, if the keycloak url is:Keycloak URL: https://one-3gk1k.worker-01-use1.prod.ataccama.link/auth/Base URL:https://one-3gk1k.worker-01-use1.prod.ataccama.link/The Valid Redirect URI will be:https://one-3gk1k.worker
Getting started with using Manta integration in Ataccama
If you just configured the integration between Manta and Ataccama and have no experience using it, here’s your guide to getting started. Integrating Data Sources First, you’ll want to turn on the integration for the Data Sources in your Ataccama Knowledge Catalog.If you are creating a new data source in ONE, simply mark the checkbox Enable Manta Integration in the Connection details. This will automatically propagate the new source to Manta. If you want to integrate an existing source in ONE, go to Source >> Connection >> Edit and mark the Enable Manta Integration checkbox and Save. To confirm that the Connection is successfully propagated to Manta, check that it appears in the Manta Admin UI (Connections >> Databases). You might have some data sources connected directly to Manta - it’s fine to keep them alongside the ones propagated from Ataccama. Note that when you trigger Calculate Lineage from the Ataccama ONE app, the Manta workflow will execute lineage revision
RDM Web application Role
I(Using admin creds) have assigned a user the role of “RDM-Work_DataEntry” from Keycloak for being able to make an entry on RDM Web app. But he is not able to see any Table or View details on the web app, Can anyone help me with the mistake i might be committing. Which role should be assigned for the purpose of being able to add/remove/delete/edit entries on RDM tables on Web App/UI?Also, Where can i find the descriptions for all the roles that are available on the Keycloak to be assigned?
Ataccama ONE Desktop - SFTP wildcard
Hello,I want to use “SFTP Download File” and use a wildcard for the filename and a date. Is that possible ? For example :/<mydirectory>/input/MYFILE_<TODAY_YYMMDD>_*.txt When i enter an asterisk symbol in the “File or Directory” section, I get an error. Thanks,Rene Requenes
Strange error during RDM publish with OnPublish
Hello community,I’m working on an RDM project which has the OnPublish implementation. However, I am encountering an error whenever I try to publish a record. The error message from the Action History states:The response message was:null and response error was 'errorMsg:Cannot get credentials for request.'I have checked all the database connections and the connection to the RDM web application on the runtime server, and there does not seem to be any issue in that regard.Is anyone familiar with this error and could provide assistance? Thank you in advance.
How to resolve ONE Desktop error “The model project is not licensed for use on this machine”Best Practice
When you use in ONE Desktop application the project synchronization with Git, you can in some cases meet with the error “The model project is not licensed for use on this machine”, followed by warning “Project signature is invalid”.At first glance, it seems there must be an issue with the license. However, mostly it is not the case, and the issue is related to Windows line endings, as Eclipse utilizes Unix syntax line endings.How to resolve this?Follow the steps below:In ONE Desktop select your project name (as an example further we use the project Reference Data Manager - Banking example), and click in the top menu panel to choose:File => Convert Line Delimiters To => Unix. Uncheck Show only text files option and click Deselect All. Open .metadata folder in the "Select files to convert" menu, select meta.mmd and manifest.xml files. Click on the small arrow next to the .metadata folder to display contents, and click on the checkbox beside the libs folder to select all
RDM Incremental load to CONFIRM state
Hello, communityI want to implement an RDM solution where we periodically load the data from a database and import them to RDM in the published state. Each load could contain new and updated records.So I set the RDM Importer step to load data to CONFIRM state in the INCREMENTAL fashion (which as description state: new records are imported, existing records not present in the input file are untouched, matched records are updated).However, I encountered an error indicating the “Key” is already exist when execute the plan.Could anyone help me on this please? I’m not sure the incremental option only works for loading to EDIT state or do I need to setup anything additionally in the column section.Thank you in advance
Import records with username to RDM
Hello Ataccama community 😀Has anyone previously used the USERNAME column type in the RDM Importer or RDM History Importer? I want to specify the [ Username ] on RDM, indicating who made changes to each record in an automated fashion.However, I got an error telling me that unknown column 'username'; I tried using “USERNAME” instead of “username,” but still had no luck.Could anyone familiar with this help me please?
Manage Nodes Access, Access Level in v.14.2 (VS v.13.9)
Hello everyone! Foreword and problem statementThere are several subcategories of terms. I need to set up access for manager and steward of each subcategory so that:The steward of certain subcategory is able to create new terms, edit old ones in his subcategory only. However, he can not publish any changes The manager, in addition to above, can publish changes, but also in his own subcategory only.I have the list of questions concerning access management in v.14.2:Is it possible to grant ownership to node? And if it’s yes, how can I implement it? Do you know any way to solve my manager-steward task? (there are my understanding of the solution below) In States of approval workflowwhat are the “User roles” : governance or identity provider ones? (looks like the second ones, but doesn’t work) what is the “OWNER” for this version of workflow?_______________________There is a long prelude to where the questions came from. Version: 13.9There is the ability to create a special capability to b
Inquiry about RDM Merge Importer Step
Hello community,I am trying to implement the RDM database synchronization. However, the data from a database does not seem to be loaded back to RDM; thus, I checked the system_input plan, which is the plan that is supposed to load data back to the RDM. The key step is the RDM Merge Importer that loads data back, but after I ran the plan, there does not seem to be any data loaded to the RDM in any viewing perspective.I also notice that the column rdm_primary_key is null for every record not in RDM prior to entering the Importer step. I’m not sure if this is an issue either.Does anyone familiar with the RDM Merge Importer step? Could you please guide me on how to use this step properly?Thank you in advance 😀
RDM reverse synchronization not working
Hello community,I’m implementing the RDM synchronization to sync data with a database. However, the data seems to sync only one way, which is from the RDM to the database. Whenever I add an additional record to the database manually, it gets removed after performing synchronization, and they don’t appear in the RDM inputs view.I have “Use Reverse Synchronization” enabled and followed the steps from Configuring RDM Synchronization with External Databases - Ataccama ONE Gen2 Platform Latest.If anyone knows how to get this working, could you please guide me on this?
Extracting data quality results- deatiled failed records
Hi,We are using Ataccama one 12.6 version. We have a project containing 40+ models with 90 + rules. I am looking for an options to get the detailed invalid/failed records for each model. Like all the attributes from the model along with invalid rule instance names.The idea is to get the whole failed dataset to automate the remediation process. We already have automated process to remediate the data using other technologies.I don’t want to use the download results in csv option, because we are transitioning to latest version of Ataccama.I am looking for couple of optionsExtract the detailed failed records from backend tables using sqls (where Ataccama stores the results) or would like to understand how those csv files are generated in backend and use the same way to get the data and store in db. Reading data through Ataccama inbuilt API’s and loading into a db. I am going through the GraphQL options. Could you please let me know, If there is way to get the similar data through GraphQL
What is the best mechanism or approach to share the desktop html profile results?
Hello everyone,I created a profile of my catalog item using ONE desktop and I would like the output(.html) to be accessible to business analyst with read access. The html is only accessible on my local computer. Where’s the best / secure location to store output reports of profiling? Please chime in with any thoughts. Thank you! 😀
Attribute Name Duplication Error for SQL Catalog Item Query
Hi, I’ve been trying to create a SQL catalog item in the UI using the query below. It’s basically joining the same table twice. The query ran successfully in IDE but it’s giving me an error: Attribute '3' is duplicated. Please make sure that attributes have unique names. Can anyone help?select x.1, x.2, ax.3 as SSN, aze.3 as EIDfrom table1 xinner join table 2 ax on (x.a = ax.b) and trim(ax.c) in ('1','S','Social Security Number US')left join table 2 aze on (x.a = aze.b) and trim(aze.c) in ('E','EMPL_PROVIDED_ID','Employee ID')where x.last_update_date > '2023-02-08 00:00:00.000'
Comparison of two columns for matches in One Desktop
Hello everyone!I have an interesting task related to the desktop, we have a table combined with certain attributes and with full names, full names are divided into columns (firstname, lastname, midname) and they need to be compared with other columns (firstname1, lastname1, midname1) it is necessary to output non-matching full names. I can't compare one table between columns, I need your help. Do you know how to implement this?I will be very grateful for your help
Have you discovered Ataccama resources?
Hi community! Have you discovered our resources section yet? 👀We always strive to bring you the latest news and develop features and ways for you to achieve the best data platform management. In this post, we’d like to give you a quick overview of the available resources and what opportunities are available for you there. Webinars 🧑💻 Find on-demand recordings of version releases, feature trainings, and live demonstrations of customer use cases to panel discussions with industry experts. Blogs 📑 Keep up to date with the latest industry news and research, and learn more about best practices and tips & tricks articles on DQ, MDM, DG, and more. News 🗞️ Be the first one to find out about Ataccama news. Success Stories 💪 Access more than twenty case studies and use cases to learn more and get inspired by different Ataccama customers from various industries implementing our products. Demos 🎬 Need a live walk-through of one of our services? You can find it all under our dem
Kafka Streaming Use Case ExamplesBest Practice
Use Case: Send records from Kafka to Ataccama IDE for parsing and other transformations and send this data to database tables Beginning set up for Kafka with MDM: ⦁ The first thing you would need to do is add the Kafka configuration to the runtime configuration file, for usage in the server The configuration should be in this format as in the picture above, where you provide the connection string, key, trustore, and keystore passwords You will also need to add this information into the serverConfig file as well, naming the server as you would like between the string tags Once the file configurations are ready for you to connect to Kafka, you can then start adding your streaming information into the model explorer The first step would be to right-click on the STREAMING button in the model explorer From there you would need to click New Consumer You would then click Kafka Stream Source, to create a new kafka streaming source and configuration Above is the screen you should see
Let's shape the future of Ataccama together
See new features and prototypes before everyone else, and have an opportunity to shape the future of new products.
Login to the community
Enter your username or e-mail address. We'll send you an e-mail with instructions to reset your password.