Skip to content

Commit 9622a97

Browse files
temporarily comment out tests segfaulting on julia v1.12 (#631)
* remove tests * fix docs * fix cuda buildkite
1 parent c181a10 commit 9622a97

4 files changed

Lines changed: 62 additions & 68 deletions

File tree

.github/workflows/multidocs.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ jobs:
1717
- uses: actions/checkout@v6
1818
- uses: julia-actions/setup-julia@v2
1919
with:
20-
version: '1'
20+
version: '1.12.2' # set to '1' after 1.12.4 release, see https://github.com/JuliaLang/Pkg.jl/pull/4568
2121
- uses: julia-actions/cache@v2
2222
# Build individual docs
2323
- run: julia --project=GNNGraphs/docs/ -e 'using Pkg; Pkg.instantiate()'

GraphNeuralNetworks/docs/src/dev.md

Lines changed: 4 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -74,22 +74,13 @@ Each PR should update the version number in the Porject.toml file of each involv
7474
the compat bounds, e.g. GraphNeuralNetworks might require a newer version of GNNGraphs.
7575

7676
## Generate Documentation Locally
77-
For generating the documentation locally
77+
Each package has its own documentation folder, e.g. `GNNGraphs/docs`. To generate the docs locally, run the following command from the root of the repository:
78+
7879
```
79-
cd docs
80-
julia
80+
# example for GNNGraphs
81+
julia --project=GNNGraphs/docs GNNGraphs/docs/make.jl
8182
```
82-
```julia
83-
(@v1.10) pkg> activate .
84-
Activating project at `~/.julia/dev/GraphNeuralNetworks/docs`
8583

86-
(docs) pkg> dev ../ ../GNNGraphs/
87-
Resolving package versions...
88-
No Changes to `~/.julia/dev/GraphNeuralNetworks/docs/Project.toml`
89-
No Changes to `~/.julia/dev/GraphNeuralNetworks/docs/Manifest.toml`
90-
91-
julia> include("make.jl")
92-
```
9384
## Benchmarking
9485

9586
You can benchmark the effect on performance of your commits using the script `perf/perf.jl`.

GraphNeuralNetworks/test/layers/conv.jl

Lines changed: 55 additions & 53 deletions
Original file line numberDiff line numberDiff line change
@@ -365,35 +365,36 @@ end
365365
end
366366
end
367367

368-
@testitem "CGConv" setup=[TolSnippet, TestModule] begin
369-
using .TestModule
370-
371-
edim = 10
372-
l = CGConv((D_IN, edim) => D_OUT, tanh, residual = false, bias = true)
373-
for g in TEST_GRAPHS
374-
g = GNNGraph(g, edata = rand(Float32, edim, g.num_edges))
375-
@test size(l(g, g.x, g.e)) == (D_OUT, g.num_nodes)
376-
test_gradients(l, g, g.x, g.e, rtol = RTOL_HIGH)
377-
end
378-
379-
# no edge features
380-
l1 = CGConv(D_IN => D_OUT, tanh, residual = false, bias = true)
381-
g1 = TEST_GRAPHS[1]
382-
@test l1(g1, g1.ndata.x) == l1(g1).ndata.x
383-
@test l1(g1, g1.ndata.x, nothing) == l1(g1).ndata.x
384-
end
385-
386-
@testitem "CGConv GPU" setup=[TolSnippet, TestModule] tags=[:gpu] begin
387-
using .TestModule
388-
edim = 10
389-
l = CGConv((D_IN, edim) => D_OUT, tanh, residual = false, bias = true)
390-
for g in TEST_GRAPHS
391-
g.graph isa AbstractSparseMatrix && continue
392-
g = GNNGraph(g, edata = rand(Float32, edim, g.num_edges))
393-
@test size(l(g, g.x, g.e)) == (D_OUT, g.num_nodes)
394-
test_gradients(l, g, g.x, g.e, rtol = RTOL_HIGH, test_gpu = true, compare_finite_diff = false)
395-
end
396-
end
368+
## TODO segfault on julia v1.12
369+
# @testitem "CGConv" setup=[TolSnippet, TestModule] begin
370+
# using .TestModule
371+
372+
# edim = 10
373+
# l = CGConv((D_IN, edim) => D_OUT, tanh, residual = false, bias = true)
374+
# for g in TEST_GRAPHS
375+
# g = GNNGraph(g, edata = rand(Float32, edim, g.num_edges))
376+
# @test size(l(g, g.x, g.e)) == (D_OUT, g.num_nodes)
377+
# test_gradients(l, g, g.x, g.e, rtol = RTOL_HIGH)
378+
# end
379+
380+
# # no edge features
381+
# l1 = CGConv(D_IN => D_OUT, tanh, residual = false, bias = true)
382+
# g1 = TEST_GRAPHS[1]
383+
# @test l1(g1, g1.ndata.x) == l1(g1).ndata.x
384+
# @test l1(g1, g1.ndata.x, nothing) == l1(g1).ndata.x
385+
# end
386+
387+
# @testitem "CGConv GPU" setup=[TolSnippet, TestModule] tags=[:gpu] begin
388+
# using .TestModule
389+
# edim = 10
390+
# l = CGConv((D_IN, edim) => D_OUT, tanh, residual = false, bias = true)
391+
# for g in TEST_GRAPHS
392+
# g.graph isa AbstractSparseMatrix && continue
393+
# g = GNNGraph(g, edata = rand(Float32, edim, g.num_edges))
394+
# @test size(l(g, g.x, g.e)) == (D_OUT, g.num_nodes)
395+
# test_gradients(l, g, g.x, g.e, rtol = RTOL_HIGH, test_gpu = true, compare_finite_diff = false)
396+
# end
397+
# end
397398

398399
@testitem "AGNNConv" setup=[TolSnippet, TestModule] begin
399400
using .TestModule
@@ -457,30 +458,31 @@ end
457458
end
458459
end
459460

460-
@testitem "GMMConv" setup=[TolSnippet, TestModule] begin
461-
using .TestModule
462-
ein_channel = 10
463-
K = 5
464-
l = GMMConv((D_IN, ein_channel) => D_OUT, K = K)
465-
for g in TEST_GRAPHS
466-
g = GNNGraph(g, edata = rand(Float32, ein_channel, g.num_edges))
467-
y = l(g, g.x, g.e)
468-
test_gradients(l, g, g.x, g.e, rtol = RTOL_HIGH)
469-
end
470-
end
471-
472-
@testitem "GMMConv GPU" setup=[TolSnippet, TestModule] tags=[:gpu] begin
473-
using .TestModule
474-
ein_channel = 10
475-
K = 5
476-
l = GMMConv((D_IN, ein_channel) => D_OUT, K = K)
477-
for g in TEST_GRAPHS
478-
g.graph isa AbstractSparseMatrix && continue
479-
g = GNNGraph(g, edata = rand(Float32, ein_channel, g.num_edges))
480-
y = l(g, g.x, g.e)
481-
test_gradients(l, g, g.x, g.e, rtol = RTOL_HIGH, test_gpu = true, compare_finite_diff = false)
482-
end
483-
end
461+
## TODO segfault on julia v1.12
462+
# @testitem "GMMConv" setup=[TolSnippet, TestModule] begin
463+
# using .TestModule
464+
# ein_channel = 10
465+
# K = 5
466+
# l = GMMConv((D_IN, ein_channel) => D_OUT, K = K)
467+
# for g in TEST_GRAPHS
468+
# g = GNNGraph(g, edata = rand(Float32, ein_channel, g.num_edges))
469+
# y = l(g, g.x, g.e)
470+
# test_gradients(l, g, g.x, g.e, rtol = RTOL_HIGH)
471+
# end
472+
# end
473+
474+
# @testitem "GMMConv GPU" setup=[TolSnippet, TestModule] tags=[:gpu] begin
475+
# using .TestModule
476+
# ein_channel = 10
477+
# K = 5
478+
# l = GMMConv((D_IN, ein_channel) => D_OUT, K = K)
479+
# for g in TEST_GRAPHS
480+
# g.graph isa AbstractSparseMatrix && continue
481+
# g = GNNGraph(g, edata = rand(Float32, ein_channel, g.num_edges))
482+
# y = l(g, g.x, g.e)
483+
# test_gradients(l, g, g.x, g.e, rtol = RTOL_HIGH, test_gpu = true, compare_finite_diff = false)
484+
# end
485+
# end
484486

485487
@testitem "SGConv" setup=[TolSnippet, TestModule] begin
486488
using .TestModule

Project.toml

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,2 +1,3 @@
11
[workspace]
2-
projects = ["GNNGraphs", "GNNlib", "GraphNeuralNetworks", "GNNLux"]
2+
projects = ["docs", "GNNGraphs", "GNNlib", "GraphNeuralNetworks", "GNNLux"]
3+

0 commit comments

Comments
 (0)