Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
22 changes: 20 additions & 2 deletions src/vtlengine/Operators/Join.py
Original file line number Diff line number Diff line change
Expand Up @@ -406,7 +406,20 @@ def execute(cls, dataset: Dataset, op: Any, left: str, right: str) -> Dataset:
return op.evaluate(left_dataset, right_dataset)

@classmethod
def validate(cls, dataset: Dataset, child: Any, op_map: Dict[str, Any]) -> None:
def validate(cls, dataset: Dataset, child: Any, op_map: Dict[str, Any]) -> Dataset:
if isinstance(child, list):
for c in child:
dataset = cls.validate(dataset, c, op_map)
else:
cls._check_bin_expr(dataset, child, op_map)
left_dataset = cls.create_dataset("left", child.left.value, dataset)
right_dataset = cls.create_dataset("right", child.right.value, dataset)
dataset, _ = cls.get_common_components(left_dataset, right_dataset)

return dataset

@classmethod
def _check_bin_expr(cls, dataset: Dataset, child: Any, op_map: Dict[str, Any]) -> None:
if not isinstance(child, BinOp):
raise Exception(
f"Invalid expression {child} on apply operator. Only BinOp are accepted"
Expand Down Expand Up @@ -438,7 +451,12 @@ def create_dataset(cls, name: str, prefix: str, dataset: Dataset) -> Dataset:
for component in dataset.components.values()
if component.name.startswith(prefix) or component.role is Role.IDENTIFIER
}
data = dataset.data[list(components.keys())] if dataset.data is not None else pd.DataFrame()
comp_names = list(components.keys())
data = (
dataset.data[comp_names]
if dataset.data is not None
else pd.DataFrame(columns=comp_names)
)

for component in components.values():
component.name = (
Expand Down
21 changes: 21 additions & 0 deletions tests/Semantic/data/DataStructure/input/GH_609-1.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,21 @@
{
"datasets": [
{
"name": "DS_1",
"DataStructure": [
{
"name": "Id_1",
"type": "Integer",
"role": "Identifier",
"nullable": false
},
{
"name": "Me_1",
"type": "Number",
"role": "Measure",
"nullable": true
}
]
}
]
}
21 changes: 21 additions & 0 deletions tests/Semantic/data/DataStructure/output/GH_609-1.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,21 @@
{
"datasets": [
{
"name": "DS_r",
"DataStructure": [
{
"name": "Id_1",
"type": "Integer",
"role": "Identifier",
"nullable": false
},
{
"name": "Me_1",
"type": "Number",
"role": "Measure",
"nullable": true
}
]
}
]
}
1 change: 1 addition & 0 deletions tests/Semantic/data/vtl/GH_609.vtl
Original file line number Diff line number Diff line change
@@ -0,0 +1 @@
DS_r := inner_join (DS_1 as d1, DS_1 as d2 apply d1 + d2);
13 changes: 13 additions & 0 deletions tests/Semantic/test_semantic.py
Original file line number Diff line number Diff line change
Expand Up @@ -1854,6 +1854,19 @@ def test_20(self):

self.BaseTest(code=code, number_inputs=number_inputs, references_names=references_names)

def test_GH_609(self):
"""
Dataset --> Dataset
Status:
Expression: DS_r := inner_join (DS_1 as d1, DS_1 as d2 apply d1 + d2);
Description: Apply operation fails on semantic run
"""
code = "GH_609"
number_inputs = 1
references_names = ["1"]

self.BaseTest(code=code, number_inputs=number_inputs, references_names=references_names)


class AggregateTests(SemanticHelper):
""" """
Expand Down
Loading