https://kedro.org/ logo
Join the conversationJoin Discord
Channels
advanced-need-help
announcements
beginners-need-help
introductions
job-posting
plugins-integrations
random
resources
welcome
Powered by Linen
advanced-need-help
  • n

    noklam

    08/24/2022, 3:11 PM
    Can u upgrade kedro-viz?
  • s

    skuma458

    08/24/2022, 3:22 PM
    have done that too, have viz version 5.0.1
  • d

    datajoely

    08/24/2022, 3:34 PM
    Are you using a virtual environment here?
  • s

    Scott H

    08/24/2022, 7:26 PM
    Is there a way to run a DAG so that one or more nodes failing in the middle don't stop the whole pipeline? I'd like it to continue running the nodes that do not depend on the failed nodes, possibly tracking the failed and skipped nodes and throwing an error or log message at the end summarizing what was failed or skipped. I've thought of writing a custom runner for this, but it occurred to me something others may have worked on or been interested in.
    a
    • 2
    • 4
  • n

    noklam

    08/24/2022, 7:30 PM
    On_node_error hook may help here, but may need something more to figure out the dag dependencies
  • s

    skuma458

    08/25/2022, 6:48 AM
    code running in a container
  • d

    datajoely

    08/25/2022, 6:49 AM
    And does a normal kedro run work ok?
  • d

    datajoely

    08/25/2022, 6:49 AM
    And are you actually using HDFS or spark?
  • s

    skuma458

    08/25/2022, 6:54 AM
    kedro pipelines are running without issue, running on spark
    d
    • 2
    • 16
  • m

    Matthias Roels

    08/25/2022, 8:52 AM
    Quick question: looking at the source code of kedro, I noticed this line: https://github.com/kedro-org/kedro/blob/0.17.7/kedro/framework/session/session.py#L179. Does that mean that in orde to use kedro, you always need it to be part of a git repo?
  • d

    datajoely

    08/25/2022, 9:18 AM
    @skuma458 no if you use the exp tracking functionality we track the git sha but don't fail without it
  • m

    Matthias Roels

    08/25/2022, 9:28 AM
    Well, it does fail when I use notebooks (I don't use the exp tracking functionality)
  • m

    Matthias Roels

    08/25/2022, 9:28 AM
    Btw: I think you replied to the wrong person :p
  • d

    datajoely

    08/25/2022, 10:28 AM
    Yes I did!
  • d

    datajoely

    08/25/2022, 10:29 AM
    That being said we may have removed the dependency on git later than 0.17.7
  • g

    gui42

    08/25/2022, 4:52 PM
    Hey folks. How would I run a piece of code whenever I run kedro? I'm thinking of running alembic migrations that way. Is there a proper hook for that?
  • d

    datajoely

    08/25/2022, 4:54 PM
    So am I right in saying the process has nothing to do with your kedro catalog etc and is more about the same time?
  • g

    gui42

    08/25/2022, 4:55 PM
    yep. More like defining schemas and adding columns to historical tables. I just want the process to be seamless and live in the same repo 😅
  • d

    datajoely

    08/25/2022, 4:55 PM
    Gotcha
  • d

    datajoely

    08/25/2022, 4:56 PM
    So I'd say there is some merit in orchestrating this with something else like Airflow or Argo just because it will be easier to debug isolated issues.... But after_pipeline_run feels like the safest place to do this
  • g

    gui42

    08/25/2022, 4:59 PM
    got it. Makes sense! I'm just mimicking the pattern for running migrations whenever you start a server in web dev. For me, it makes sense whenever you had long-lasting pipelines that run in a production setting.
  • t

    ToniMaroni

    08/28/2022, 10:18 AM
    Hi Kedro team, I'm really enjoying structuring my projects with Kedro, thank you very much for this great piece of software! I'm curious about the search functionality in kedro viz experiment section. I would like to be able to search in my experiments based on a value of a variable that I am tracking, but it seems that the search is only going thru the experiment metadata. Is there a specific syntax to search thru the tracked data? If not is there a way to programatically add "notes" an experiment? Thanks for the feedback!
  • d

    datajoely

    08/28/2022, 10:32 AM
    So not yet but this is super valuable feedback
  • d

    datajoely

    08/28/2022, 10:33 AM
    Could you raise an issue on the Kedro Viz repo?
  • m

    Matthias Roels

    08/29/2022, 7:42 AM
    Nope: it is still there in 0.18.x https://github.com/kedro-org/kedro/blob/main/kedro/framework/session/session.py#L157. This means that whenever I am not in a repository, I can't even start a kedro session...
  • d

    datajoely

    08/29/2022, 8:30 AM
    What error are you getting we should be catching it?
  • b

    beats-like-a-helix

    08/29/2022, 7:49 PM
    Let's say I have a modular execution pipeline which is supposed to use a trained model to analyse N datasets. I understand how to create N instances of my modular pipeline (one for each dataset), but I was hoping there was some way to auto-generate catalog entries for these datasets at runtime. ie. I would have
    dataset_1_result, dataset_2_result, ..., dataset_N_result
    in the catalog without having to write them in the catalog manually. I understand it's just yaml input at the end of the day, and I could just write a hacky script to populate the catalog with these entires, but just wondered if there's a more "Kedronic" way to accomplish this. Cheers!
  • d

    datajoely

    08/29/2022, 7:50 PM
    Before pipeline run hook!
  • b

    beats-like-a-helix

    08/29/2022, 7:51 PM
    Awesome, not actually got up to speed with hooks yet. Much obliged!
  • a

    antheas

    08/30/2022, 6:27 AM
    after_catalog_created hook with after_context_created to get the context. I did the same With before pipeline run you won't be able to access those datasets with say kedro ipython. Therefore there's no point in saving them But if you use my way you'll have to namespace your datasets uniquely
Powered by Linen
Title
a

antheas

08/30/2022, 6:27 AM
after_catalog_created hook with after_context_created to get the context. I did the same With before pipeline run you won't be able to access those datasets with say kedro ipython. Therefore there's no point in saving them But if you use my way you'll have to namespace your datasets uniquely
View count: 1